278 lines
7.1 KiB
C
278 lines
7.1 KiB
C
/*
|
|
* Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved.
|
|
* Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without modification,
|
|
* are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright notice, this list of
|
|
* conditions and the following disclaimer.
|
|
*
|
|
* 2. Redistributions in binary form must reproduce the above copyright notice, this list
|
|
* of conditions and the following disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* 3. Neither the name of the copyright holder nor the names of its contributors may be used
|
|
* to endorse or promote products derived from this software without specific prior written
|
|
* permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
|
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR
|
|
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
|
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
|
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
|
|
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
|
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
|
|
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
|
|
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "pthread.h"
|
|
#include "pprivate.h"
|
|
|
|
|
|
int pthread_attr_init(pthread_attr_t *attr)
|
|
{
|
|
if (attr == NULL) {
|
|
return EINVAL;
|
|
}
|
|
|
|
attr->detachstate = PTHREAD_CREATE_JOINABLE;
|
|
attr->schedpolicy = SCHED_RR;
|
|
attr->schedparam.sched_priority = LOSCFG_BASE_CORE_TSK_DEFAULT_PRIO;
|
|
attr->inheritsched = PTHREAD_INHERIT_SCHED;
|
|
attr->scope = PTHREAD_SCOPE_PROCESS;
|
|
attr->stackaddr_set = 0;
|
|
attr->stackaddr = NULL;
|
|
attr->stacksize_set = 1;
|
|
attr->stacksize = LOSCFG_BASE_CORE_TSK_DEFAULT_STACK_SIZE;
|
|
|
|
#if (LOSCFG_KERNEL_SMP == YES)
|
|
attr->cpuset.__bits[0] = 0;
|
|
#endif
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_destroy(pthread_attr_t *attr)
|
|
{
|
|
if (attr == NULL) {
|
|
return EINVAL;
|
|
}
|
|
|
|
/* Nothing to do here... */
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_setdetachstate(pthread_attr_t *attr, int detachState)
|
|
{
|
|
if ((attr != NULL) && ((detachState == PTHREAD_CREATE_JOINABLE) || (detachState == PTHREAD_CREATE_DETACHED))) {
|
|
attr->detachstate = (UINT32)detachState;
|
|
return ENOERR;
|
|
}
|
|
|
|
return EINVAL;
|
|
}
|
|
|
|
int pthread_attr_getdetachstate(const pthread_attr_t *attr, int *detachState)
|
|
{
|
|
if ((attr == NULL) || (detachState == NULL)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*detachState = (int)attr->detachstate;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_setscope(pthread_attr_t *attr, int scope)
|
|
{
|
|
if (attr == NULL) {
|
|
return EINVAL;
|
|
}
|
|
|
|
if (scope == PTHREAD_SCOPE_PROCESS) {
|
|
attr->scope = (unsigned int)scope;
|
|
return ENOERR;
|
|
}
|
|
|
|
if (scope == PTHREAD_SCOPE_SYSTEM) {
|
|
return ENOTSUP;
|
|
}
|
|
|
|
return EINVAL;
|
|
}
|
|
|
|
int pthread_attr_getscope(const pthread_attr_t *attr, int *scope)
|
|
{
|
|
if ((attr == NULL) || (scope == NULL)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*scope = (int)attr->scope;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_setinheritsched(pthread_attr_t *attr, int inherit)
|
|
{
|
|
if ((attr != NULL) && ((inherit == PTHREAD_INHERIT_SCHED) || (inherit == PTHREAD_EXPLICIT_SCHED))) {
|
|
attr->inheritsched = (UINT32)inherit;
|
|
return ENOERR;
|
|
}
|
|
|
|
return EINVAL;
|
|
}
|
|
|
|
int pthread_attr_getinheritsched(const pthread_attr_t *attr, int *inherit)
|
|
{
|
|
if ((attr == NULL) || (inherit == NULL)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*inherit = (int)attr->inheritsched;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_setschedpolicy(pthread_attr_t *attr, int policy)
|
|
{
|
|
if ((attr != NULL) && (policy == SCHED_RR)) {
|
|
attr->schedpolicy = SCHED_RR;
|
|
return ENOERR;
|
|
}
|
|
|
|
return EINVAL;
|
|
}
|
|
|
|
int pthread_attr_getschedpolicy(const pthread_attr_t *attr, int *policy)
|
|
{
|
|
if ((attr == NULL) || (policy == NULL)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*policy = (int)attr->schedpolicy;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_setschedparam(pthread_attr_t *attr, const struct sched_param *param)
|
|
{
|
|
if ((attr == NULL) || (param == NULL)) {
|
|
return EINVAL;
|
|
} else if ((param->sched_priority < 0) || (param->sched_priority > OS_TASK_PRIORITY_LOWEST)) {
|
|
return ENOTSUP;
|
|
}
|
|
|
|
attr->schedparam = *param;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_getschedparam(const pthread_attr_t *attr, struct sched_param *param)
|
|
{
|
|
if ((attr == NULL) || (param == NULL)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*param = attr->schedparam;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
/*
|
|
* Set starting address of stack. Whether this is at the start or end of
|
|
* the memory block allocated for the stack depends on whether the stack
|
|
* grows up or down.
|
|
*/
|
|
int pthread_attr_setstackaddr(pthread_attr_t *attr, void *stackAddr)
|
|
{
|
|
if (attr == NULL) {
|
|
return EINVAL;
|
|
}
|
|
|
|
attr->stackaddr_set = 1;
|
|
attr->stackaddr = stackAddr;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_getstackaddr(const pthread_attr_t *attr, void **stackAddr)
|
|
{
|
|
if (((attr != NULL) && (stackAddr != NULL)) && attr->stackaddr_set) {
|
|
*stackAddr = attr->stackaddr;
|
|
return ENOERR;
|
|
}
|
|
|
|
return EINVAL; /* Stack address not set, return EINVAL. */
|
|
}
|
|
|
|
int pthread_attr_setstacksize(pthread_attr_t *attr, size_t stackSize)
|
|
{
|
|
/* Reject inadequate stack sizes */
|
|
if ((attr == NULL) || (stackSize < PTHREAD_STACK_MIN)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
attr->stacksize_set = 1;
|
|
attr->stacksize = stackSize;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
int pthread_attr_getstacksize(const pthread_attr_t *attr, size_t *stackSize)
|
|
{
|
|
/* Reject attempts to get a stack size when one has not been set. */
|
|
if ((attr == NULL) || (stackSize == NULL) || (!attr->stacksize_set)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*stackSize = attr->stacksize;
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
/*
|
|
* Set the cpu affinity mask
|
|
*/
|
|
int pthread_attr_setaffinity_np(pthread_attr_t* attr, size_t cpusetsize, const cpu_set_t* cpuset)
|
|
{
|
|
#if (LOSCFG_KERNEL_SMP == YES)
|
|
if (attr == NULL) {
|
|
return EINVAL;
|
|
}
|
|
|
|
if ((cpuset == NULL) || (cpusetsize == 0)) {
|
|
attr->cpuset.__bits[0] = 0;
|
|
return ENOERR;
|
|
}
|
|
|
|
if ((cpusetsize != sizeof(cpu_set_t)) || (cpuset->__bits[0] > LOSCFG_KERNEL_CPU_MASK)) {
|
|
return EINVAL;
|
|
}
|
|
|
|
attr->cpuset = *cpuset;
|
|
#endif
|
|
|
|
return ENOERR;
|
|
}
|
|
|
|
/*
|
|
* Get the cpu affinity mask
|
|
*/
|
|
int pthread_attr_getaffinity_np(const pthread_attr_t* attr, size_t cpusetsize, cpu_set_t* cpuset)
|
|
{
|
|
#if (LOSCFG_KERNEL_SMP == YES)
|
|
if ((attr == NULL) || (cpuset == NULL) || (cpusetsize != sizeof(cpu_set_t))) {
|
|
return EINVAL;
|
|
}
|
|
|
|
*cpuset = attr->cpuset;
|
|
#endif
|
|
|
|
return ENOERR;
|
|
}
|
|
|