Ai
3 Star 1 Fork 0

Gitee 极速下载/apache-apr

加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
文件
此仓库是为了提升国内下载速度的镜像仓库,每日同步一次。 原始仓库: https://github.com/apache/apr
克隆/下载
proc_mutex.c 48.84 KB
一键复制 编辑 原始数据 按行查看 历史
Eric Covener 提交于 2024-08-21 21:01 +08:00 . Use APR_TIMEUP for ETIMEDOUT too
1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699
/* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "apr.h"
#include "apr_strings.h"
#include "apr_arch_proc_mutex.h"
#include "apr_arch_file_io.h" /* for apr_mkstemp() */
#include "apr_md5.h" /* for apr_md5() */
#include "apr_atomic.h"
APR_DECLARE(apr_status_t) apr_proc_mutex_destroy(apr_proc_mutex_t *mutex)
{
apr_status_t rv = apr_proc_mutex_cleanup(mutex);
if (rv == APR_SUCCESS) {
apr_pool_cleanup_kill(mutex->pool, mutex, apr_proc_mutex_cleanup);
}
return rv;
}
#if APR_HAS_POSIXSEM_SERIALIZE || APR_HAS_FCNTL_SERIALIZE || \
APR_HAS_SYSVSEM_SERIALIZE
static apr_status_t proc_mutex_no_child_init(apr_proc_mutex_t **mutex,
apr_pool_t *cont,
const char *fname)
{
return APR_SUCCESS;
}
#endif
#if APR_HAS_POSIXSEM_SERIALIZE || APR_HAS_PROC_PTHREAD_SERIALIZE
static apr_status_t proc_mutex_no_perms_set(apr_proc_mutex_t *mutex,
apr_fileperms_t perms,
apr_uid_t uid,
apr_gid_t gid)
{
return APR_ENOTIMPL;
}
#endif
#if APR_HAS_FCNTL_SERIALIZE \
|| APR_HAS_FLOCK_SERIALIZE \
|| (APR_HAS_SYSVSEM_SERIALIZE \
&& !defined(HAVE_SEMTIMEDOP)) \
|| (APR_HAS_POSIXSEM_SERIALIZE \
&& !defined(HAVE_SEM_TIMEDWAIT)) \
|| (APR_HAS_PROC_PTHREAD_SERIALIZE \
&& !defined(HAVE_PTHREAD_MUTEX_TIMEDLOCK) \
&& !defined(HAVE_PTHREAD_CONDATTR_SETPSHARED))
static apr_status_t proc_mutex_spinsleep_timedacquire(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
#define SLEEP_TIME apr_time_from_msec(10)
apr_status_t rv;
for (;;) {
rv = apr_proc_mutex_trylock(mutex);
if (!APR_STATUS_IS_EBUSY(rv)) {
if (rv == APR_SUCCESS) {
mutex->curr_locked = 1;
}
break;
}
if (timeout <= 0) {
rv = APR_TIMEUP;
break;
}
if (timeout > SLEEP_TIME) {
apr_sleep(SLEEP_TIME);
timeout -= SLEEP_TIME;
}
else {
apr_sleep(timeout);
timeout = 0;
}
}
return rv;
}
#endif
#if APR_HAS_POSIXSEM_SERIALIZE
#ifndef SEM_FAILED
#define SEM_FAILED (-1)
#endif
static apr_status_t proc_mutex_posix_cleanup(void *mutex_)
{
apr_proc_mutex_t *mutex = mutex_;
if (sem_close(mutex->os.psem_interproc) < 0) {
return errno;
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_posix_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
#define APR_POSIXSEM_NAME_MAX 30
#define APR_POSIXSEM_NAME_MIN 13
sem_t *psem;
char semname[APR_MD5_DIGESTSIZE * 2 + 2];
/*
* This bogusness is to follow what appears to be the
* lowest common denominator in Posix semaphore naming:
* - start with '/'
* - be at most 14 chars
* - be unique and not match anything on the filesystem
*
* Because of this, we use fname to generate an md5 hex checksum
* and use that as the name of the semaphore. If no filename was
* given, we create one based on the time. We tuck the name
* away, since it might be useful for debugging.
*
* To make this as robust as possible, we initially try something
* larger (and hopefully more unique) and gracefully fail down to the
* LCD above.
*
* NOTE: Darwin (Mac OS X) seems to be the most restrictive
* implementation. Versions previous to Darwin 6.2 had the 14
* char limit, but later rev's allow up to 31 characters.
*
*/
if (fname) {
unsigned char digest[APR_MD5_DIGESTSIZE]; /* note dependency on semname here */
const char *hex = "0123456789abcdef";
char *p = semname;
int i;
apr_md5(digest, fname, strlen(fname));
*p++ = '/'; /* must start with /, right? */
for (i = 0; i < sizeof(digest); i++) {
*p++ = hex[digest[i] >> 4];
*p++ = hex[digest[i] & 0xF];
}
semname[APR_POSIXSEM_NAME_MAX] = '\0';
} else {
apr_time_t now;
unsigned long sec;
unsigned long usec;
now = apr_time_now();
sec = apr_time_sec(now);
usec = apr_time_usec(now);
apr_snprintf(semname, sizeof(semname), "/ApR.%lxZ%lx", sec, usec);
}
do {
psem = sem_open(semname, O_CREAT | O_EXCL, 0644, 1);
} while (psem == (sem_t *)SEM_FAILED && errno == EINTR);
if (psem == (sem_t *)SEM_FAILED) {
if (errno == ENAMETOOLONG) {
/* Oh well, good try */
semname[APR_POSIXSEM_NAME_MIN] = '\0';
} else {
return errno;
}
do {
psem = sem_open(semname, O_CREAT | O_EXCL, 0644, 1);
} while (psem == (sem_t *)SEM_FAILED && errno == EINTR);
}
if (psem == (sem_t *)SEM_FAILED) {
return errno;
}
/* Ahhh. The joys of Posix sems. Predelete it... */
sem_unlink(semname);
new_mutex->os.psem_interproc = psem;
new_mutex->fname = apr_pstrdup(new_mutex->pool, semname);
apr_pool_cleanup_register(new_mutex->pool, (void *)new_mutex,
apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
return APR_SUCCESS;
}
static apr_status_t proc_mutex_posix_acquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = sem_wait(mutex->os.psem_interproc);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_posix_tryacquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = sem_trywait(mutex->os.psem_interproc);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
if (errno == EAGAIN) {
return APR_EBUSY;
}
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
#if defined(HAVE_SEM_TIMEDWAIT)
static apr_status_t proc_mutex_posix_timedacquire(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
if (timeout <= 0) {
apr_status_t rv = proc_mutex_posix_tryacquire(mutex);
return (rv == APR_EBUSY) ? APR_TIMEUP : rv;
}
else {
int rc;
struct timespec abstime;
timeout += apr_time_now();
abstime.tv_sec = apr_time_sec(timeout);
abstime.tv_nsec = apr_time_usec(timeout) * 1000; /* nanoseconds */
do {
rc = sem_timedwait(mutex->os.psem_interproc, &abstime);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
if (errno == ETIMEDOUT) {
return APR_TIMEUP;
}
return errno;
}
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
#endif
static apr_status_t proc_mutex_posix_release(apr_proc_mutex_t *mutex)
{
mutex->curr_locked = 0;
if (sem_post(mutex->os.psem_interproc) < 0) {
/* any failure is probably fatal, so no big deal to leave
* ->curr_locked at 0. */
return errno;
}
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_posixsem_methods =
{
#if APR_PROCESS_LOCK_IS_GLOBAL || !APR_HAS_THREADS || defined(POSIXSEM_IS_GLOBAL)
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
#else
0,
#endif
proc_mutex_posix_create,
proc_mutex_posix_acquire,
proc_mutex_posix_tryacquire,
#if defined(HAVE_SEM_TIMEDWAIT)
proc_mutex_posix_timedacquire,
#else
proc_mutex_spinsleep_timedacquire,
#endif
proc_mutex_posix_release,
proc_mutex_posix_cleanup,
proc_mutex_no_child_init,
proc_mutex_no_perms_set,
APR_LOCK_POSIXSEM,
"posixsem"
};
#endif /* Posix sem implementation */
#if APR_HAS_SYSVSEM_SERIALIZE
static struct sembuf proc_mutex_op_on;
static struct sembuf proc_mutex_op_try;
static struct sembuf proc_mutex_op_off;
static void proc_mutex_sysv_setup(void)
{
proc_mutex_op_on.sem_num = 0;
proc_mutex_op_on.sem_op = -1;
proc_mutex_op_on.sem_flg = SEM_UNDO;
proc_mutex_op_try.sem_num = 0;
proc_mutex_op_try.sem_op = -1;
proc_mutex_op_try.sem_flg = SEM_UNDO | IPC_NOWAIT;
proc_mutex_op_off.sem_num = 0;
proc_mutex_op_off.sem_op = 1;
proc_mutex_op_off.sem_flg = SEM_UNDO;
}
static apr_status_t proc_mutex_sysv_cleanup(void *mutex_)
{
apr_proc_mutex_t *mutex=mutex_;
union semun ick;
if (mutex->os.crossproc != -1) {
ick.val = 0;
semctl(mutex->os.crossproc, 0, IPC_RMID, ick);
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_sysv_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
union semun ick;
apr_status_t rv;
new_mutex->os.crossproc = semget(IPC_PRIVATE, 1, IPC_CREAT | 0600);
if (new_mutex->os.crossproc == -1) {
rv = errno;
proc_mutex_sysv_cleanup(new_mutex);
return rv;
}
ick.val = 1;
if (semctl(new_mutex->os.crossproc, 0, SETVAL, ick) < 0) {
rv = errno;
proc_mutex_sysv_cleanup(new_mutex);
new_mutex->os.crossproc = -1;
return rv;
}
new_mutex->curr_locked = 0;
apr_pool_cleanup_register(new_mutex->pool,
(void *)new_mutex, apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
return APR_SUCCESS;
}
static apr_status_t proc_mutex_sysv_acquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = semop(mutex->os.crossproc, &proc_mutex_op_on, 1);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_sysv_tryacquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = semop(mutex->os.crossproc, &proc_mutex_op_try, 1);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
if (errno == EAGAIN) {
return APR_EBUSY;
}
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
#if defined(HAVE_SEMTIMEDOP)
static apr_status_t proc_mutex_sysv_timedacquire(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
if (timeout <= 0) {
apr_status_t rv = proc_mutex_sysv_tryacquire(mutex);
return (rv == APR_EBUSY) ? APR_TIMEUP : rv;
}
else {
int rc;
struct timespec reltime;
reltime.tv_sec = apr_time_sec(timeout);
reltime.tv_nsec = apr_time_usec(timeout) * 1000; /* nanoseconds */
do {
rc = semtimedop(mutex->os.crossproc, &proc_mutex_op_on, 1,
&reltime);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
if (errno == EAGAIN || errno == ETIMEDOUT) {
return APR_TIMEUP;
}
return errno;
}
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
#endif
static apr_status_t proc_mutex_sysv_release(apr_proc_mutex_t *mutex)
{
int rc;
mutex->curr_locked = 0;
do {
rc = semop(mutex->os.crossproc, &proc_mutex_op_off, 1);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_sysv_perms_set(apr_proc_mutex_t *mutex,
apr_fileperms_t perms,
apr_uid_t uid,
apr_gid_t gid)
{
union semun ick;
struct semid_ds buf;
buf.sem_perm.uid = uid;
buf.sem_perm.gid = gid;
buf.sem_perm.mode = apr_unix_perms2mode(perms);
ick.buf = &buf;
if (semctl(mutex->os.crossproc, 0, IPC_SET, ick) < 0) {
return errno;
}
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_sysv_methods =
{
#if APR_PROCESS_LOCK_IS_GLOBAL || !APR_HAS_THREADS || defined(SYSVSEM_IS_GLOBAL)
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
#else
0,
#endif
proc_mutex_sysv_create,
proc_mutex_sysv_acquire,
proc_mutex_sysv_tryacquire,
#if defined(HAVE_SEMTIMEDOP)
proc_mutex_sysv_timedacquire,
#else
proc_mutex_spinsleep_timedacquire,
#endif
proc_mutex_sysv_release,
proc_mutex_sysv_cleanup,
proc_mutex_no_child_init,
proc_mutex_sysv_perms_set,
APR_LOCK_SYSVSEM,
"sysvsem"
};
#endif /* SysV sem implementation */
#if APR_HAS_PROC_PTHREAD_SERIALIZE
#ifndef APR_USE_PROC_PTHREAD_MUTEX_COND
#if defined(HAVE_PTHREAD_CONDATTR_SETPSHARED) \
&& !defined(HAVE_PTHREAD_MUTEX_TIMEDLOCK)
#define APR_USE_PROC_PTHREAD_MUTEX_COND 1
#else
#define APR_USE_PROC_PTHREAD_MUTEX_COND 0
#endif
#endif
/* The mmap()ed pthread_interproc is the native pthread_mutex_t followed
* by a refcounter to track children using it. We want to avoid calling
* pthread_mutex_destroy() on the shared mutex area while it is in use by
* another process, because this may mark the shared pthread_mutex_t as
* invalid for everyone, including forked children (unlike "sysvsem" for
* example), causing unexpected errors or deadlocks (PR 49504). So the
* last process (parent or child) referencing the mutex will effectively
* destroy it.
*/
typedef struct {
#define proc_pthread_cast(m) \
((proc_pthread_mutex_t *)(m)->os.pthread_interproc)
pthread_mutex_t mutex;
#define proc_pthread_mutex(m) \
(proc_pthread_cast(m)->mutex)
#if APR_USE_PROC_PTHREAD_MUTEX_COND
pthread_cond_t cond;
#define proc_pthread_mutex_cond(m) \
(proc_pthread_cast(m)->cond)
apr_int32_t cond_locked;
#define proc_pthread_mutex_cond_locked(m) \
(proc_pthread_cast(m)->cond_locked)
apr_uint32_t cond_num_waiters;
#define proc_pthread_mutex_cond_num_waiters(m) \
(proc_pthread_cast(m)->cond_num_waiters)
#define proc_pthread_mutex_is_cond(m) \
((m)->pthread_refcounting && proc_pthread_mutex_cond_locked(m) != -1)
#endif /* APR_USE_PROC_PTHREAD_MUTEX_COND */
apr_uint32_t refcount;
#define proc_pthread_mutex_refcount(m) \
(proc_pthread_cast(m)->refcount)
} proc_pthread_mutex_t;
static APR_INLINE int proc_pthread_mutex_inc(apr_proc_mutex_t *mutex)
{
if (mutex->pthread_refcounting) {
apr_atomic_inc32(&proc_pthread_mutex_refcount(mutex));
return 1;
}
return 0;
}
static APR_INLINE int proc_pthread_mutex_dec(apr_proc_mutex_t *mutex)
{
if (mutex->pthread_refcounting) {
return apr_atomic_dec32(&proc_pthread_mutex_refcount(mutex));
}
return 0;
}
static apr_status_t proc_pthread_mutex_unref(void *mutex_)
{
apr_proc_mutex_t *mutex=mutex_;
apr_status_t rv;
#if APR_USE_PROC_PTHREAD_MUTEX_COND
if (proc_pthread_mutex_is_cond(mutex)) {
mutex->curr_locked = 0;
}
else
#endif /* APR_USE_PROC_PTHREAD_MUTEX_COND */
if (mutex->curr_locked == 1) {
if ((rv = pthread_mutex_unlock(&proc_pthread_mutex(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
return rv;
}
}
if (!proc_pthread_mutex_dec(mutex)) {
#if APR_USE_PROC_PTHREAD_MUTEX_COND
if (proc_pthread_mutex_is_cond(mutex) &&
(rv = pthread_cond_destroy(&proc_pthread_mutex_cond(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
return rv;
}
#endif /* APR_USE_PROC_PTHREAD_MUTEX_COND */
if ((rv = pthread_mutex_destroy(&proc_pthread_mutex(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
return rv;
}
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_pthread_cleanup(void *mutex_)
{
apr_proc_mutex_t *mutex=mutex_;
apr_status_t rv;
/* curr_locked is set to -1 until the mutex has been created */
if (mutex->curr_locked != -1) {
if ((rv = proc_pthread_mutex_unref(mutex))) {
return rv;
}
}
if (munmap(mutex->os.pthread_interproc, sizeof(proc_pthread_mutex_t))) {
return errno;
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_pthread_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
apr_status_t rv;
int fd;
pthread_mutexattr_t mattr;
fd = open("/dev/zero", O_RDWR);
if (fd < 0) {
return errno;
}
new_mutex->os.pthread_interproc = mmap(NULL, sizeof(proc_pthread_mutex_t),
PROT_READ | PROT_WRITE, MAP_SHARED,
fd, 0);
if (new_mutex->os.pthread_interproc == MAP_FAILED) {
new_mutex->os.pthread_interproc = NULL;
rv = errno;
close(fd);
return rv;
}
close(fd);
new_mutex->pthread_refcounting = 1;
new_mutex->curr_locked = -1; /* until the mutex has been created */
#if APR_USE_PROC_PTHREAD_MUTEX_COND
proc_pthread_mutex_cond_locked(new_mutex) = -1;
#endif
if ((rv = pthread_mutexattr_init(&mattr))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
return rv;
}
if ((rv = pthread_mutexattr_setpshared(&mattr, PTHREAD_PROCESS_SHARED))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
pthread_mutexattr_destroy(&mattr);
return rv;
}
#if defined(HAVE_PTHREAD_MUTEX_ROBUST) || defined(HAVE_PTHREAD_MUTEX_ROBUST_NP)
#ifdef HAVE_PTHREAD_MUTEX_ROBUST
rv = pthread_mutexattr_setrobust(&mattr, PTHREAD_MUTEX_ROBUST);
#else
rv = pthread_mutexattr_setrobust_np(&mattr, PTHREAD_MUTEX_ROBUST_NP);
#endif
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
pthread_mutexattr_destroy(&mattr);
return rv;
}
if ((rv = pthread_mutexattr_setprotocol(&mattr, PTHREAD_PRIO_INHERIT))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
pthread_mutexattr_destroy(&mattr);
return rv;
}
#endif /* HAVE_PTHREAD_MUTEX_ROBUST[_NP] */
#if defined(APR_THREAD_DEBUG)
/* ignore errors. */
if ((rv = pthread_mutexattr_settype(&mattr, PTHREAD_MUTEX_ERRORCHECK))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
pthread_mutexattr_destroy(&mattr);
return rv;
}
#endif
if ((rv = pthread_mutex_init(&proc_pthread_mutex(new_mutex), &mattr))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
pthread_mutexattr_destroy(&mattr);
return rv;
}
proc_pthread_mutex_refcount(new_mutex) = 1; /* first/parent reference */
new_mutex->curr_locked = 0; /* mutex created now */
if ((rv = pthread_mutexattr_destroy(&mattr))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
proc_mutex_pthread_cleanup(new_mutex);
return rv;
}
apr_pool_cleanup_register(new_mutex->pool,
(void *)new_mutex,
apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
return APR_SUCCESS;
}
static apr_status_t proc_mutex_pthread_child_init(apr_proc_mutex_t **mutex,
apr_pool_t *pool,
const char *fname)
{
(*mutex)->curr_locked = 0;
if (proc_pthread_mutex_inc(*mutex)) {
apr_pool_cleanup_register(pool, *mutex, proc_pthread_mutex_unref,
apr_pool_cleanup_null);
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_pthread_acquire_ex(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
apr_status_t rv;
#if APR_USE_PROC_PTHREAD_MUTEX_COND
if (proc_pthread_mutex_is_cond(mutex)) {
if ((rv = pthread_mutex_lock(&proc_pthread_mutex(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
#if defined(HAVE_PTHREAD_MUTEX_ROBUST) || defined(HAVE_PTHREAD_MUTEX_ROBUST_NP)
/* Okay, our owner died. Let's try to make it consistent again. */
if (rv == EOWNERDEAD) {
proc_pthread_mutex_dec(mutex);
#ifdef HAVE_PTHREAD_MUTEX_ROBUST
pthread_mutex_consistent(&proc_pthread_mutex(mutex));
#else
pthread_mutex_consistent_np(&proc_pthread_mutex(mutex));
#endif
}
else
#endif
return rv;
}
if (!proc_pthread_mutex_cond_locked(mutex)) {
rv = APR_SUCCESS;
}
else if (!timeout) {
rv = APR_TIMEUP;
}
else {
struct timespec abstime;
if (timeout > 0) {
timeout += apr_time_now();
abstime.tv_sec = apr_time_sec(timeout);
abstime.tv_nsec = apr_time_usec(timeout) * 1000; /* nanoseconds */
}
proc_pthread_mutex_cond_num_waiters(mutex)++;
do {
if (timeout < 0) {
rv = pthread_cond_wait(&proc_pthread_mutex_cond(mutex),
&proc_pthread_mutex(mutex));
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
break;
}
}
else {
rv = pthread_cond_timedwait(&proc_pthread_mutex_cond(mutex),
&proc_pthread_mutex(mutex),
&abstime);
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
if (rv == ETIMEDOUT) {
rv = APR_TIMEUP;
}
break;
}
}
} while (proc_pthread_mutex_cond_locked(mutex));
proc_pthread_mutex_cond_num_waiters(mutex)--;
}
if (rv != APR_SUCCESS) {
pthread_mutex_unlock(&proc_pthread_mutex(mutex));
return rv;
}
proc_pthread_mutex_cond_locked(mutex) = 1;
rv = pthread_mutex_unlock(&proc_pthread_mutex(mutex));
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
return rv;
}
}
else
#endif /* APR_USE_PROC_PTHREAD_MUTEX_COND */
{
if (timeout < 0) {
rv = pthread_mutex_lock(&proc_pthread_mutex(mutex));
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
}
}
else if (!timeout) {
rv = pthread_mutex_trylock(&proc_pthread_mutex(mutex));
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
if (rv == EBUSY) {
return APR_TIMEUP;
}
}
}
else
#if defined(HAVE_PTHREAD_MUTEX_TIMEDLOCK)
{
struct timespec abstime;
timeout += apr_time_now();
abstime.tv_sec = apr_time_sec(timeout);
abstime.tv_nsec = apr_time_usec(timeout) * 1000; /* nanoseconds */
rv = pthread_mutex_timedlock(&proc_pthread_mutex(mutex), &abstime);
if (rv) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
if (rv == ETIMEDOUT) {
return APR_TIMEUP;
}
}
}
if (rv) {
#if defined(HAVE_PTHREAD_MUTEX_ROBUST) || defined(HAVE_PTHREAD_MUTEX_ROBUST_NP)
/* Okay, our owner died. Let's try to make it consistent again. */
if (rv == EOWNERDEAD) {
proc_pthread_mutex_dec(mutex);
#ifdef HAVE_PTHREAD_MUTEX_ROBUST
pthread_mutex_consistent(&proc_pthread_mutex(mutex));
#else
pthread_mutex_consistent_np(&proc_pthread_mutex(mutex));
#endif
}
else
#endif
return rv;
}
#else /* !HAVE_PTHREAD_MUTEX_TIMEDLOCK */
return proc_mutex_spinsleep_timedacquire(mutex, timeout);
#endif
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_pthread_acquire(apr_proc_mutex_t *mutex)
{
return proc_mutex_pthread_acquire_ex(mutex, -1);
}
static apr_status_t proc_mutex_pthread_tryacquire(apr_proc_mutex_t *mutex)
{
apr_status_t rv = proc_mutex_pthread_acquire_ex(mutex, 0);
return (rv == APR_TIMEUP) ? APR_EBUSY : rv;
}
static apr_status_t proc_mutex_pthread_timedacquire(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
return proc_mutex_pthread_acquire_ex(mutex, (timeout <= 0) ? 0 : timeout);
}
static apr_status_t proc_mutex_pthread_release(apr_proc_mutex_t *mutex)
{
apr_status_t rv;
#if APR_USE_PROC_PTHREAD_MUTEX_COND
if (proc_pthread_mutex_is_cond(mutex)) {
if ((rv = pthread_mutex_lock(&proc_pthread_mutex(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
#if defined(HAVE_PTHREAD_MUTEX_ROBUST) || defined(HAVE_PTHREAD_MUTEX_ROBUST_NP)
/* Okay, our owner died. Let's try to make it consistent again. */
if (rv == EOWNERDEAD) {
proc_pthread_mutex_dec(mutex);
#ifdef HAVE_PTHREAD_MUTEX_ROBUST
pthread_mutex_consistent(&proc_pthread_mutex(mutex));
#else
pthread_mutex_consistent_np(&proc_pthread_mutex(mutex));
#endif
}
else
#endif
return rv;
}
if (!proc_pthread_mutex_cond_locked(mutex)) {
rv = APR_EINVAL;
}
else if (!proc_pthread_mutex_cond_num_waiters(mutex)) {
rv = APR_SUCCESS;
}
else {
rv = pthread_cond_signal(&proc_pthread_mutex_cond(mutex));
#ifdef HAVE_ZOS_PTHREADS
if (rv) {
rv = errno;
}
#endif
}
if (rv != APR_SUCCESS) {
pthread_mutex_unlock(&proc_pthread_mutex(mutex));
return rv;
}
proc_pthread_mutex_cond_locked(mutex) = 0;
}
#endif /* APR_USE_PROC_PTHREAD_MUTEX_COND */
mutex->curr_locked = 0;
if ((rv = pthread_mutex_unlock(&proc_pthread_mutex(mutex)))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
return rv;
}
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_proc_pthread_methods =
{
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
proc_mutex_pthread_create,
proc_mutex_pthread_acquire,
proc_mutex_pthread_tryacquire,
proc_mutex_pthread_timedacquire,
proc_mutex_pthread_release,
proc_mutex_pthread_cleanup,
proc_mutex_pthread_child_init,
proc_mutex_no_perms_set,
APR_LOCK_PROC_PTHREAD,
"pthread"
};
#if APR_USE_PROC_PTHREAD_MUTEX_COND
static apr_status_t proc_mutex_pthread_cond_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
apr_status_t rv;
pthread_condattr_t cattr;
rv = proc_mutex_pthread_create(new_mutex, fname);
if (rv != APR_SUCCESS) {
return rv;
}
if ((rv = pthread_condattr_init(&cattr))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
apr_pool_cleanup_run(new_mutex->pool, new_mutex,
apr_proc_mutex_cleanup);
return rv;
}
if ((rv = pthread_condattr_setpshared(&cattr, PTHREAD_PROCESS_SHARED))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
pthread_condattr_destroy(&cattr);
apr_pool_cleanup_run(new_mutex->pool, new_mutex,
apr_proc_mutex_cleanup);
return rv;
}
if ((rv = pthread_cond_init(&proc_pthread_mutex_cond(new_mutex),
&cattr))) {
#ifdef HAVE_ZOS_PTHREADS
rv = errno;
#endif
pthread_condattr_destroy(&cattr);
apr_pool_cleanup_run(new_mutex->pool, new_mutex,
apr_proc_mutex_cleanup);
return rv;
}
pthread_condattr_destroy(&cattr);
proc_pthread_mutex_cond_locked(new_mutex) = 0;
proc_pthread_mutex_cond_num_waiters(new_mutex) = 0;
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_proc_pthread_cond_methods =
{
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
proc_mutex_pthread_cond_create,
proc_mutex_pthread_acquire,
proc_mutex_pthread_tryacquire,
proc_mutex_pthread_timedacquire,
proc_mutex_pthread_release,
proc_mutex_pthread_cleanup,
proc_mutex_pthread_child_init,
proc_mutex_no_perms_set,
APR_LOCK_PROC_PTHREAD,
"pthread"
};
#endif
#endif
#if APR_HAS_FCNTL_SERIALIZE
static struct flock proc_mutex_lock_it;
static struct flock proc_mutex_unlock_it;
static apr_status_t proc_mutex_fcntl_release(apr_proc_mutex_t *);
static void proc_mutex_fcntl_setup(void)
{
proc_mutex_lock_it.l_whence = SEEK_SET; /* from current point */
proc_mutex_lock_it.l_start = 0; /* -"- */
proc_mutex_lock_it.l_len = 0; /* until end of file */
proc_mutex_lock_it.l_type = F_WRLCK; /* set exclusive/write lock */
proc_mutex_lock_it.l_pid = 0; /* pid not actually interesting */
proc_mutex_unlock_it.l_whence = SEEK_SET; /* from current point */
proc_mutex_unlock_it.l_start = 0; /* -"- */
proc_mutex_unlock_it.l_len = 0; /* until end of file */
proc_mutex_unlock_it.l_type = F_UNLCK; /* set exclusive/write lock */
proc_mutex_unlock_it.l_pid = 0; /* pid not actually interesting */
}
static apr_status_t proc_mutex_fcntl_cleanup(void *mutex_)
{
apr_status_t status = APR_SUCCESS;
apr_proc_mutex_t *mutex=mutex_;
if (mutex->curr_locked == 1) {
status = proc_mutex_fcntl_release(mutex);
if (status != APR_SUCCESS)
return status;
}
if (mutex->interproc) {
status = apr_file_close(mutex->interproc);
}
if (!mutex->interproc_closing
&& mutex->os.crossproc != -1
&& close(mutex->os.crossproc) == -1
&& status == APR_SUCCESS) {
status = errno;
}
return status;
}
static apr_status_t proc_mutex_fcntl_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
int rv;
if (fname) {
new_mutex->fname = apr_pstrdup(new_mutex->pool, fname);
rv = apr_file_open(&new_mutex->interproc, new_mutex->fname,
APR_FOPEN_CREATE | APR_FOPEN_WRITE | APR_FOPEN_EXCL,
APR_FPROT_UREAD | APR_FPROT_UWRITE | APR_FPROT_GREAD | APR_FPROT_WREAD,
new_mutex->pool);
}
else {
new_mutex->fname = apr_pstrdup(new_mutex->pool, "/tmp/aprXXXXXX");
rv = apr_file_mktemp(&new_mutex->interproc, new_mutex->fname,
APR_FOPEN_CREATE | APR_FOPEN_WRITE | APR_FOPEN_EXCL,
new_mutex->pool);
}
if (rv != APR_SUCCESS) {
return rv;
}
new_mutex->os.crossproc = new_mutex->interproc->filedes;
new_mutex->interproc_closing = 1;
new_mutex->curr_locked = 0;
unlink(new_mutex->fname);
apr_pool_cleanup_register(new_mutex->pool,
(void*)new_mutex,
apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
return APR_SUCCESS;
}
static apr_status_t proc_mutex_fcntl_acquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = fcntl(mutex->os.crossproc, F_SETLKW, &proc_mutex_lock_it);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
mutex->curr_locked=1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_fcntl_tryacquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = fcntl(mutex->os.crossproc, F_SETLK, &proc_mutex_lock_it);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
#if FCNTL_TRYACQUIRE_EACCES
if (errno == EACCES) {
#else
if (errno == EAGAIN) {
#endif
return APR_EBUSY;
}
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_fcntl_release(apr_proc_mutex_t *mutex)
{
int rc;
mutex->curr_locked=0;
do {
rc = fcntl(mutex->os.crossproc, F_SETLKW, &proc_mutex_unlock_it);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_fcntl_perms_set(apr_proc_mutex_t *mutex,
apr_fileperms_t perms,
apr_uid_t uid,
apr_gid_t gid)
{
if (mutex->fname) {
if (!(perms & APR_FPROT_GSETID))
gid = -1;
if (fchown(mutex->os.crossproc, uid, gid) < 0) {
return errno;
}
}
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_fcntl_methods =
{
#if APR_PROCESS_LOCK_IS_GLOBAL || !APR_HAS_THREADS || defined(FCNTL_IS_GLOBAL)
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
#else
0,
#endif
proc_mutex_fcntl_create,
proc_mutex_fcntl_acquire,
proc_mutex_fcntl_tryacquire,
proc_mutex_spinsleep_timedacquire,
proc_mutex_fcntl_release,
proc_mutex_fcntl_cleanup,
proc_mutex_no_child_init,
proc_mutex_fcntl_perms_set,
APR_LOCK_FCNTL,
"fcntl"
};
#endif /* fcntl implementation */
#if APR_HAS_FLOCK_SERIALIZE
static apr_status_t proc_mutex_flock_release(apr_proc_mutex_t *);
static apr_status_t proc_mutex_flock_cleanup(void *mutex_)
{
apr_status_t status = APR_SUCCESS;
apr_proc_mutex_t *mutex=mutex_;
if (mutex->curr_locked == 1) {
status = proc_mutex_flock_release(mutex);
if (status != APR_SUCCESS)
return status;
}
if (mutex->interproc) { /* if it was opened properly */
status = apr_file_close(mutex->interproc);
}
if (!mutex->interproc_closing
&& mutex->os.crossproc != -1
&& close(mutex->os.crossproc) == -1
&& status == APR_SUCCESS) {
status = errno;
}
if (mutex->fname) {
unlink(mutex->fname);
}
return status;
}
static apr_status_t proc_mutex_flock_create(apr_proc_mutex_t *new_mutex,
const char *fname)
{
int rv;
if (fname) {
new_mutex->fname = apr_pstrdup(new_mutex->pool, fname);
rv = apr_file_open(&new_mutex->interproc, new_mutex->fname,
APR_FOPEN_CREATE | APR_FOPEN_WRITE | APR_FOPEN_EXCL,
APR_FPROT_UREAD | APR_FPROT_UWRITE,
new_mutex->pool);
}
else {
new_mutex->fname = apr_pstrdup(new_mutex->pool, "/tmp/aprXXXXXX");
rv = apr_file_mktemp(&new_mutex->interproc, new_mutex->fname,
APR_FOPEN_CREATE | APR_FOPEN_WRITE | APR_FOPEN_EXCL,
new_mutex->pool);
}
if (rv != APR_SUCCESS) {
proc_mutex_flock_cleanup(new_mutex);
return rv;
}
new_mutex->os.crossproc = new_mutex->interproc->filedes;
new_mutex->interproc_closing = 1;
new_mutex->curr_locked = 0;
apr_pool_cleanup_register(new_mutex->pool, (void *)new_mutex,
apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
return APR_SUCCESS;
}
static apr_status_t proc_mutex_flock_acquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = flock(mutex->os.crossproc, LOCK_EX);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_flock_tryacquire(apr_proc_mutex_t *mutex)
{
int rc;
do {
rc = flock(mutex->os.crossproc, LOCK_EX | LOCK_NB);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
if (errno == EWOULDBLOCK || errno == EAGAIN) {
return APR_EBUSY;
}
return errno;
}
mutex->curr_locked = 1;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_flock_release(apr_proc_mutex_t *mutex)
{
int rc;
mutex->curr_locked = 0;
do {
rc = flock(mutex->os.crossproc, LOCK_UN);
} while (rc < 0 && errno == EINTR);
if (rc < 0) {
return errno;
}
return APR_SUCCESS;
}
static apr_status_t proc_mutex_flock_child_init(apr_proc_mutex_t **mutex,
apr_pool_t *pool,
const char *fname)
{
apr_proc_mutex_t *new_mutex;
int rv;
if (!fname) {
fname = (*mutex)->fname;
if (!fname) {
return APR_SUCCESS;
}
}
new_mutex = (apr_proc_mutex_t *)apr_pmemdup(pool, *mutex,
sizeof(apr_proc_mutex_t));
new_mutex->pool = pool;
new_mutex->fname = apr_pstrdup(pool, fname);
rv = apr_file_open(&new_mutex->interproc, new_mutex->fname,
APR_FOPEN_WRITE, 0, new_mutex->pool);
if (rv != APR_SUCCESS) {
return rv;
}
new_mutex->os.crossproc = new_mutex->interproc->filedes;
new_mutex->interproc_closing = 1;
*mutex = new_mutex;
return APR_SUCCESS;
}
static apr_status_t proc_mutex_flock_perms_set(apr_proc_mutex_t *mutex,
apr_fileperms_t perms,
apr_uid_t uid,
apr_gid_t gid)
{
if (mutex->fname) {
if (!(perms & APR_FPROT_GSETID))
gid = -1;
if (fchown(mutex->os.crossproc, uid, gid) < 0) {
return errno;
}
}
return APR_SUCCESS;
}
static const apr_proc_mutex_unix_lock_methods_t mutex_flock_methods =
{
#if APR_PROCESS_LOCK_IS_GLOBAL || !APR_HAS_THREADS || defined(FLOCK_IS_GLOBAL)
APR_PROCESS_LOCK_MECH_IS_GLOBAL,
#else
0,
#endif
proc_mutex_flock_create,
proc_mutex_flock_acquire,
proc_mutex_flock_tryacquire,
proc_mutex_spinsleep_timedacquire,
proc_mutex_flock_release,
proc_mutex_flock_cleanup,
proc_mutex_flock_child_init,
proc_mutex_flock_perms_set,
APR_LOCK_FLOCK,
"flock"
};
#endif /* flock implementation */
void apr_proc_mutex_unix_setup_lock(void)
{
/* setup only needed for sysvsem and fnctl */
#if APR_HAS_SYSVSEM_SERIALIZE
proc_mutex_sysv_setup();
#endif
#if APR_HAS_FCNTL_SERIALIZE
proc_mutex_fcntl_setup();
#endif
}
static apr_status_t proc_mutex_choose_method(apr_proc_mutex_t *new_mutex,
apr_lockmech_e mech,
apr_os_proc_mutex_t *ospmutex)
{
#if APR_HAS_PROC_PTHREAD_SERIALIZE
new_mutex->os.pthread_interproc = NULL;
#endif
#if APR_HAS_POSIXSEM_SERIALIZE
new_mutex->os.psem_interproc = NULL;
#endif
#if APR_HAS_SYSVSEM_SERIALIZE || APR_HAS_FCNTL_SERIALIZE || APR_HAS_FLOCK_SERIALIZE
new_mutex->os.crossproc = -1;
#if APR_HAS_FCNTL_SERIALIZE || APR_HAS_FLOCK_SERIALIZE
new_mutex->interproc = NULL;
new_mutex->interproc_closing = 0;
#endif
#endif
switch (mech) {
case APR_LOCK_FCNTL:
#if APR_HAS_FCNTL_SERIALIZE
new_mutex->meth = &mutex_fcntl_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
case APR_LOCK_FLOCK:
#if APR_HAS_FLOCK_SERIALIZE
new_mutex->meth = &mutex_flock_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
case APR_LOCK_SYSVSEM:
#if APR_HAS_SYSVSEM_SERIALIZE
new_mutex->meth = &mutex_sysv_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
case APR_LOCK_POSIXSEM:
#if APR_HAS_POSIXSEM_SERIALIZE
new_mutex->meth = &mutex_posixsem_methods;
if (ospmutex) {
if (ospmutex->psem_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.psem_interproc = ospmutex->psem_interproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
case APR_LOCK_PROC_PTHREAD:
#if APR_HAS_PROC_PTHREAD_SERIALIZE
new_mutex->meth = &mutex_proc_pthread_methods;
if (ospmutex) {
if (ospmutex->pthread_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.pthread_interproc = ospmutex->pthread_interproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
case APR_LOCK_DEFAULT_TIMED:
#if APR_HAS_PROC_PTHREAD_SERIALIZE \
&& (APR_USE_PROC_PTHREAD_MUTEX_COND \
|| defined(HAVE_PTHREAD_MUTEX_TIMEDLOCK)) \
&& defined(HAVE_PTHREAD_MUTEX_ROBUST)
#if APR_USE_PROC_PTHREAD_MUTEX_COND
new_mutex->meth = &mutex_proc_pthread_cond_methods;
#else
new_mutex->meth = &mutex_proc_pthread_methods;
#endif
if (ospmutex) {
if (ospmutex->pthread_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.pthread_interproc = ospmutex->pthread_interproc;
}
break;
#elif APR_HAS_SYSVSEM_SERIALIZE && defined(HAVE_SEMTIMEDOP)
new_mutex->meth = &mutex_sysv_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
break;
#elif APR_HAS_POSIXSEM_SERIALIZE && defined(HAVE_SEM_TIMEDWAIT)
new_mutex->meth = &mutex_posixsem_methods;
if (ospmutex) {
if (ospmutex->psem_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.psem_interproc = ospmutex->psem_interproc;
}
break;
#endif
/* fall trough */
case APR_LOCK_DEFAULT:
#if APR_USE_FLOCK_SERIALIZE
new_mutex->meth = &mutex_flock_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#elif APR_USE_SYSVSEM_SERIALIZE
new_mutex->meth = &mutex_sysv_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#elif APR_USE_FCNTL_SERIALIZE
new_mutex->meth = &mutex_fcntl_methods;
if (ospmutex) {
if (ospmutex->crossproc == -1) {
return APR_EINVAL;
}
new_mutex->os.crossproc = ospmutex->crossproc;
}
#elif APR_USE_PROC_PTHREAD_SERIALIZE
new_mutex->meth = &mutex_proc_pthread_methods;
if (ospmutex) {
if (ospmutex->pthread_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.pthread_interproc = ospmutex->pthread_interproc;
}
#elif APR_USE_POSIXSEM_SERIALIZE
new_mutex->meth = &mutex_posixsem_methods;
if (ospmutex) {
if (ospmutex->psem_interproc == NULL) {
return APR_EINVAL;
}
new_mutex->os.psem_interproc = ospmutex->psem_interproc;
}
#else
return APR_ENOTIMPL;
#endif
break;
default:
return APR_ENOTIMPL;
}
return APR_SUCCESS;
}
APR_DECLARE(const char *) apr_proc_mutex_defname(void)
{
apr_proc_mutex_t mutex;
if (proc_mutex_choose_method(&mutex, APR_LOCK_DEFAULT,
NULL) != APR_SUCCESS) {
return "unknown";
}
return apr_proc_mutex_name(&mutex);
}
static apr_status_t proc_mutex_create(apr_proc_mutex_t *new_mutex, apr_lockmech_e mech, const char *fname)
{
apr_status_t rv;
if ((rv = proc_mutex_choose_method(new_mutex, mech,
NULL)) != APR_SUCCESS) {
return rv;
}
if ((rv = new_mutex->meth->create(new_mutex, fname)) != APR_SUCCESS) {
return rv;
}
return APR_SUCCESS;
}
APR_DECLARE(apr_status_t) apr_proc_mutex_create(apr_proc_mutex_t **mutex,
const char *fname,
apr_lockmech_e mech,
apr_pool_t *pool)
{
apr_proc_mutex_t *new_mutex;
apr_status_t rv;
new_mutex = apr_pcalloc(pool, sizeof(apr_proc_mutex_t));
new_mutex->pool = pool;
if ((rv = proc_mutex_create(new_mutex, mech, fname)) != APR_SUCCESS)
return rv;
*mutex = new_mutex;
return APR_SUCCESS;
}
APR_DECLARE(apr_status_t) apr_proc_mutex_child_init(apr_proc_mutex_t **mutex,
const char *fname,
apr_pool_t *pool)
{
return (*mutex)->meth->child_init(mutex, pool, fname);
}
APR_DECLARE(apr_status_t) apr_proc_mutex_lock(apr_proc_mutex_t *mutex)
{
return mutex->meth->acquire(mutex);
}
APR_DECLARE(apr_status_t) apr_proc_mutex_trylock(apr_proc_mutex_t *mutex)
{
return mutex->meth->tryacquire(mutex);
}
APR_DECLARE(apr_status_t) apr_proc_mutex_timedlock(apr_proc_mutex_t *mutex,
apr_interval_time_t timeout)
{
return mutex->meth->timedacquire(mutex, timeout);
}
APR_DECLARE(apr_status_t) apr_proc_mutex_unlock(apr_proc_mutex_t *mutex)
{
return mutex->meth->release(mutex);
}
APR_DECLARE(apr_status_t) apr_proc_mutex_cleanup(void *mutex)
{
return ((apr_proc_mutex_t *)mutex)->meth->cleanup(mutex);
}
APR_DECLARE(apr_lockmech_e) apr_proc_mutex_mech(apr_proc_mutex_t *mutex)
{
return mutex->meth->mech;
}
APR_DECLARE(const char *) apr_proc_mutex_name(apr_proc_mutex_t *mutex)
{
return mutex->meth->name;
}
APR_DECLARE(const char *) apr_proc_mutex_lockfile(apr_proc_mutex_t *mutex)
{
/* POSIX sems use the fname field but don't use a file,
* so be careful. */
#if APR_HAS_FLOCK_SERIALIZE
if (mutex->meth == &mutex_flock_methods) {
return mutex->fname;
}
#endif
#if APR_HAS_FCNTL_SERIALIZE
if (mutex->meth == &mutex_fcntl_methods) {
return mutex->fname;
}
#endif
return NULL;
}
APR_PERMS_SET_IMPLEMENT(proc_mutex)
{
apr_proc_mutex_t *mutex = (apr_proc_mutex_t *)theproc_mutex;
return mutex->meth->perms_set(mutex, perms, uid, gid);
}
APR_POOL_IMPLEMENT_ACCESSOR(proc_mutex)
/* Implement OS-specific accessors defined in apr_portable.h */
APR_DECLARE(apr_status_t) apr_os_proc_mutex_get_ex(apr_os_proc_mutex_t *ospmutex,
apr_proc_mutex_t *pmutex,
apr_lockmech_e *mech)
{
*ospmutex = pmutex->os;
if (mech) {
*mech = pmutex->meth->mech;
}
return APR_SUCCESS;
}
APR_DECLARE(apr_status_t) apr_os_proc_mutex_get(apr_os_proc_mutex_t *ospmutex,
apr_proc_mutex_t *pmutex)
{
return apr_os_proc_mutex_get_ex(ospmutex, pmutex, NULL);
}
APR_DECLARE(apr_status_t) apr_os_proc_mutex_put_ex(apr_proc_mutex_t **pmutex,
apr_os_proc_mutex_t *ospmutex,
apr_lockmech_e mech,
int register_cleanup,
apr_pool_t *pool)
{
apr_status_t rv;
if (pool == NULL) {
return APR_ENOPOOL;
}
if ((*pmutex) == NULL) {
(*pmutex) = (apr_proc_mutex_t *)apr_pcalloc(pool,
sizeof(apr_proc_mutex_t));
(*pmutex)->pool = pool;
}
rv = proc_mutex_choose_method(*pmutex, mech, ospmutex);
#if APR_HAS_FCNTL_SERIALIZE || APR_HAS_FLOCK_SERIALIZE
if (rv == APR_SUCCESS) {
rv = apr_os_file_put(&(*pmutex)->interproc, &(*pmutex)->os.crossproc,
0, pool);
}
#endif
if (rv == APR_SUCCESS && register_cleanup) {
apr_pool_cleanup_register(pool, *pmutex, apr_proc_mutex_cleanup,
apr_pool_cleanup_null);
}
return rv;
}
APR_DECLARE(apr_status_t) apr_os_proc_mutex_put(apr_proc_mutex_t **pmutex,
apr_os_proc_mutex_t *ospmutex,
apr_pool_t *pool)
{
return apr_os_proc_mutex_put_ex(pmutex, ospmutex, APR_LOCK_DEFAULT,
0, pool);
}
Loading...
马建仓 AI 助手
尝试更多
代码解读
代码找茬
代码优化
C/C++
1
https://gitee.com/mirrors/apache-apr.git
git@gitee.com:mirrors/apache-apr.git
mirrors
apache-apr
apache-apr
trunk

搜索帮助