forked from xuos/xiuos
delete task only when it's not in running.
This commit is contained in:
parent
03039cbdab
commit
a7cbb0d041
|
@ -74,7 +74,7 @@ Modification:
|
||||||
|
|
||||||
#include "cortex_a9.h"
|
#include "cortex_a9.h"
|
||||||
|
|
||||||
#define NR_CPU 3
|
#define NR_CPU 4
|
||||||
|
|
||||||
__attribute__((always_inline)) static inline uint32_t user_mode()
|
__attribute__((always_inline)) static inline uint32_t user_mode()
|
||||||
{
|
{
|
||||||
|
|
|
@ -110,50 +110,56 @@ void handle_fiq(void)
|
||||||
extern void context_switch(struct context**, struct context*);
|
extern void context_switch(struct context**, struct context*);
|
||||||
void dabort_handler(struct trapframe* r)
|
void dabort_handler(struct trapframe* r)
|
||||||
{
|
{
|
||||||
xizi_enter_kernel();
|
if (xizi_is_in_kernel()) {
|
||||||
|
uint32_t dfs, dfa;
|
||||||
uint32_t dfs, dfa;
|
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 0" : "=r"(dfs)::);
|
||||||
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 0" : "=r"(dfs)::);
|
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 0" : "=r"(dfa)::);
|
||||||
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 0" : "=r"(dfa)::);
|
|
||||||
|
|
||||||
if (r->pc < KERN_MEM_BASE) { // Exception occured in User space: exit
|
|
||||||
ERROR("dabort in user space: %s\n", cur_cpu()->task->name);
|
|
||||||
LOG("program counter: 0x%x caused\n", r->pc);
|
|
||||||
LOG("data abort at 0x%x, status 0x%x\n", dfa, dfs);
|
|
||||||
_abort_reason(dfs);
|
|
||||||
dump_tf(r);
|
|
||||||
sys_exit(cur_cpu()->task);
|
|
||||||
context_switch(&cur_cpu()->task->main_thread.context, cur_cpu()->scheduler);
|
|
||||||
} else { // Exception occured in Kernel space: panic
|
|
||||||
LOG("program counter: 0x%x caused\n", r->pc);
|
LOG("program counter: 0x%x caused\n", r->pc);
|
||||||
LOG("data abort at 0x%x, status 0x%x\n", dfa, dfs);
|
LOG("data abort at 0x%x, status 0x%x\n", dfa, dfs);
|
||||||
_abort_reason(dfs);
|
_abort_reason(dfs);
|
||||||
dump_tf(r);
|
dump_tf(r);
|
||||||
panic("data abort exception\n");
|
panic("data abort exception\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
xizi_enter_kernel();
|
||||||
|
|
||||||
|
uint32_t dfs, dfa;
|
||||||
|
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 0" : "=r"(dfs)::);
|
||||||
|
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 0" : "=r"(dfa)::);
|
||||||
|
|
||||||
|
ERROR("dabort in user space: %s\n", cur_cpu()->task->name);
|
||||||
|
LOG("program counter: 0x%x caused\n", r->pc);
|
||||||
|
LOG("data abort at 0x%x, status 0x%x\n", dfa, dfs);
|
||||||
|
_abort_reason(dfs);
|
||||||
|
dump_tf(r);
|
||||||
|
sys_exit(cur_cpu()->task);
|
||||||
|
context_switch(&cur_cpu()->task->main_thread.context, cur_cpu()->scheduler);
|
||||||
}
|
}
|
||||||
|
|
||||||
void iabort_handler(struct trapframe* r)
|
void iabort_handler(struct trapframe* r)
|
||||||
{
|
{
|
||||||
xizi_enter_kernel();
|
if (xizi_is_in_kernel()) {
|
||||||
uint32_t ifs, ifa;
|
uint32_t ifs, ifa;
|
||||||
|
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 1" : "=r"(ifs)::);
|
||||||
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 1" : "=r"(ifs)::);
|
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 2" : "=r"(ifa)::);
|
||||||
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 2" : "=r"(ifa)::);
|
LOG("program counter: 0x%x caused\n", r->pc);
|
||||||
|
|
||||||
if (r->pc < KERN_MEM_BASE) { // Exception occured in User space: exit
|
|
||||||
ERROR("iabort in user space: %s\n", cur_cpu()->task->name);
|
|
||||||
LOG("program counter: 0x%x(%s) caused\n", r->pc, cur_cpu()->task);
|
|
||||||
LOG("prefetch abort at 0x%x, status 0x%x\n", ifa, ifs);
|
|
||||||
_abort_reason(ifs);
|
|
||||||
dump_tf(r);
|
|
||||||
sys_exit(cur_cpu()->task);
|
|
||||||
context_switch(&cur_cpu()->task->main_thread.context, cur_cpu()->scheduler);
|
|
||||||
} else { // Exception occured in Kernel space: panic
|
|
||||||
LOG("program counter: 0x%x(%s) caused\n", r->pc, cur_cpu()->task);
|
|
||||||
LOG("prefetch abort at 0x%x, status 0x%x\n", ifa, ifs);
|
LOG("prefetch abort at 0x%x, status 0x%x\n", ifa, ifs);
|
||||||
_abort_reason(ifs);
|
_abort_reason(ifs);
|
||||||
dump_tf(r);
|
dump_tf(r);
|
||||||
panic("prefetch abort exception\n");
|
panic("prefetch abort exception\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
xizi_enter_kernel();
|
||||||
|
|
||||||
|
uint32_t ifs, ifa;
|
||||||
|
__asm__ __volatile__("mrc p15, 0, %0, c5, c0, 1" : "=r"(ifs)::);
|
||||||
|
__asm__ __volatile__("mrc p15, 0, %0, c6, c0, 2" : "=r"(ifa)::);
|
||||||
|
|
||||||
|
ERROR("iabort in user space: %s\n", cur_cpu()->task->name);
|
||||||
|
LOG("program counter: 0x%x(%s) caused\n", r->pc, cur_cpu()->task);
|
||||||
|
LOG("prefetch abort at 0x%x, status 0x%x\n", ifa, ifs);
|
||||||
|
_abort_reason(ifs);
|
||||||
|
dump_tf(r);
|
||||||
|
sys_exit(cur_cpu()->task);
|
||||||
|
context_switch(&cur_cpu()->task->main_thread.context, cur_cpu()->scheduler);
|
||||||
}
|
}
|
||||||
|
|
|
@ -85,9 +85,10 @@ static void _sys_irq_init(int cpu_id)
|
||||||
vector_base[5] = (uint32_t)handle_reserved; // Reserved
|
vector_base[5] = (uint32_t)handle_reserved; // Reserved
|
||||||
vector_base[6] = (uint32_t)trap_irq_enter; // IRQ
|
vector_base[6] = (uint32_t)trap_irq_enter; // IRQ
|
||||||
vector_base[7] = (uint32_t)handle_fiq; // FIQ
|
vector_base[7] = (uint32_t)handle_fiq; // FIQ
|
||||||
|
|
||||||
|
gic_init();
|
||||||
}
|
}
|
||||||
/* active hardware irq responser */
|
/* active hardware irq responser */
|
||||||
gic_init();
|
|
||||||
xizi_trap_driver.switch_hw_irqtbl((uint32_t*)&_vector_jumper);
|
xizi_trap_driver.switch_hw_irqtbl((uint32_t*)&_vector_jumper);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -139,29 +140,6 @@ static void _bind_irq_handler(int irq, irq_handler_t handler)
|
||||||
xizi_trap_driver.sw_irqtbl[irq].handler = handler;
|
xizi_trap_driver.sw_irqtbl[irq].handler = handler;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool _send_sgi(uint32_t irq, uint32_t bitmask, enum SgiFilterType type)
|
|
||||||
{
|
|
||||||
if (bitmask > (1 << NR_CPU) - 1) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
enum _gicd_sgi_filter sgi_filter;
|
|
||||||
switch (type) {
|
|
||||||
case SgiFilter_TargetList:
|
|
||||||
sgi_filter = kGicSgiFilter_UseTargetList;
|
|
||||||
break;
|
|
||||||
case SgiFilter_AllOtherCPUs:
|
|
||||||
sgi_filter = kGicSgiFilter_AllOtherCPUs;
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
sgi_filter = kGicSgiFilter_OnlyThisCPU;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
gic_send_sgi(irq, bitmask, sgi_filter);
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static uint32_t _hw_before_irq()
|
static uint32_t _hw_before_irq()
|
||||||
{
|
{
|
||||||
|
|
||||||
|
@ -217,7 +195,6 @@ static struct XiziTrapDriver xizi_trap_driver = {
|
||||||
.switch_hw_irqtbl = _switch_hw_irqtbl,
|
.switch_hw_irqtbl = _switch_hw_irqtbl,
|
||||||
|
|
||||||
.bind_irq_handler = _bind_irq_handler,
|
.bind_irq_handler = _bind_irq_handler,
|
||||||
.send_sgi = _send_sgi,
|
|
||||||
|
|
||||||
.is_interruptable = _is_interruptable,
|
.is_interruptable = _is_interruptable,
|
||||||
.hw_before_irq = _hw_before_irq,
|
.hw_before_irq = _hw_before_irq,
|
||||||
|
|
|
@ -164,24 +164,6 @@ static void _bind_irq_handler(int irq, irq_handler_t handler)
|
||||||
xizi_trap_driver.sw_irqtbl[irq].handler = handler;
|
xizi_trap_driver.sw_irqtbl[irq].handler = handler;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool _send_sgi(uint32_t irq, uint32_t bitmask, enum SgiFilterType type)
|
|
||||||
{
|
|
||||||
if (bitmask > (1 << NR_CPU) - 1) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
int cpu_id = 0;
|
|
||||||
while (bitmask != 0) {
|
|
||||||
if ((bitmask & 0x1) != 0) {
|
|
||||||
XScuGic_SoftwareIntr(&IntcInstance, irq, cpu_id);
|
|
||||||
}
|
|
||||||
cpu_id++;
|
|
||||||
bitmask >>= 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static uint32_t _hw_before_irq()
|
static uint32_t _hw_before_irq()
|
||||||
{
|
{
|
||||||
|
|
||||||
|
@ -233,7 +215,6 @@ static struct XiziTrapDriver xizi_trap_driver = {
|
||||||
.switch_hw_irqtbl = _switch_hw_irqtbl,
|
.switch_hw_irqtbl = _switch_hw_irqtbl,
|
||||||
|
|
||||||
.bind_irq_handler = _bind_irq_handler,
|
.bind_irq_handler = _bind_irq_handler,
|
||||||
.send_sgi = _send_sgi,
|
|
||||||
|
|
||||||
.is_interruptable = _is_interruptable,
|
.is_interruptable = _is_interruptable,
|
||||||
.hw_before_irq = _hw_before_irq,
|
.hw_before_irq = _hw_before_irq,
|
||||||
|
|
|
@ -66,9 +66,8 @@ struct XiziTrapDriver {
|
||||||
void (*cpu_irq_disable)();
|
void (*cpu_irq_disable)();
|
||||||
void (*single_irq_enable)(int irq, int cpu, int prio);
|
void (*single_irq_enable)(int irq, int cpu, int prio);
|
||||||
void (*single_irq_disable)(int irq, int cpu);
|
void (*single_irq_disable)(int irq, int cpu);
|
||||||
uint32_t* (*switch_hw_irqtbl)(uint32_t*);
|
|
||||||
|
|
||||||
bool (*send_sgi)(uint32_t, uint32_t, enum SgiFilterType);
|
uint32_t* (*switch_hw_irqtbl)(uint32_t*);
|
||||||
void (*bind_irq_handler)(int, irq_handler_t);
|
void (*bind_irq_handler)(int, irq_handler_t);
|
||||||
|
|
||||||
/* check if no if interruptable */
|
/* check if no if interruptable */
|
||||||
|
|
|
@ -35,7 +35,10 @@ IPC_SERVER_INTERFACE(Ipc_intr_3, 1);
|
||||||
IPC_SERVER_REGISTER_INTERFACES(IpcSwIntrHandler, 1, Ipc_intr_3);
|
IPC_SERVER_REGISTER_INTERFACES(IpcSwIntrHandler, 1, Ipc_intr_3);
|
||||||
int main()
|
int main()
|
||||||
{
|
{
|
||||||
register_irq(SW_INTERRUPT_3, Ipc_intr_3);
|
if (register_irq(SW_INTERRUPT_3, Ipc_intr_3) == -1) {
|
||||||
|
printf("TEST_SW_HDLR: bind failed");
|
||||||
|
exit();
|
||||||
|
}
|
||||||
ipc_server_loop(&IpcSwIntrHandler);
|
ipc_server_loop(&IpcSwIntrHandler);
|
||||||
|
|
||||||
exit();
|
exit();
|
||||||
|
|
|
@ -51,3 +51,4 @@ struct spinlock whole_kernel_lock;
|
||||||
|
|
||||||
void xizi_enter_kernel();
|
void xizi_enter_kernel();
|
||||||
void xizi_leave_kernel();
|
void xizi_leave_kernel();
|
||||||
|
bool xizi_is_in_kernel();
|
|
@ -93,4 +93,6 @@ int sys_state(sys_state_option option, sys_state_info* info);
|
||||||
int sys_mmap(uintptr_t vaddr, uintptr_t paddr, int len, int is_dev);
|
int sys_mmap(uintptr_t vaddr, uintptr_t paddr, int len, int is_dev);
|
||||||
|
|
||||||
int sys_register_irq(int irq_num, int irq_opcode);
|
int sys_register_irq(int irq_num, int irq_opcode);
|
||||||
|
int sys_unbind_irq_all(struct TaskMicroDescriptor* task);
|
||||||
|
int sys_unbind_irq(struct TaskMicroDescriptor* task, int irq_num);
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -64,6 +64,8 @@ struct Thread {
|
||||||
struct TaskMicroDescriptor {
|
struct TaskMicroDescriptor {
|
||||||
/* task debug resources */
|
/* task debug resources */
|
||||||
int pid;
|
int pid;
|
||||||
|
bool bind_irq;
|
||||||
|
bool dead;
|
||||||
char name[TASK_NAME_MAX_LEN];
|
char name[TASK_NAME_MAX_LEN];
|
||||||
|
|
||||||
/// @todo support return value
|
/// @todo support return value
|
||||||
|
|
|
@ -175,9 +175,10 @@ int task_exec(struct TaskMicroDescriptor* task, char* img_start, char* name, cha
|
||||||
}
|
}
|
||||||
strncpy(task->name, last, sizeof(task->name));
|
strncpy(task->name, last, sizeof(task->name));
|
||||||
|
|
||||||
xizi_pager.free_user_pgdir(&task->pgdir);
|
if (task->pgdir.pd_addr != NULL) {
|
||||||
|
xizi_pager.free_user_pgdir(&task->pgdir);
|
||||||
|
}
|
||||||
task->pgdir = pgdir;
|
task->pgdir = pgdir;
|
||||||
|
|
||||||
task->heap_base = ALIGNUP(load_size, PAGE_SIZE);
|
task->heap_base = ALIGNUP(load_size, PAGE_SIZE);
|
||||||
task->mem_size = task->heap_base + USER_STACK_SIZE;
|
task->mem_size = task->heap_base + USER_STACK_SIZE;
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -39,51 +39,7 @@ Modification:
|
||||||
int sys_exit(struct TaskMicroDescriptor* ptask)
|
int sys_exit(struct TaskMicroDescriptor* ptask)
|
||||||
{
|
{
|
||||||
assert(ptask != NULL);
|
assert(ptask != NULL);
|
||||||
|
ptask->dead = true;
|
||||||
/* handle sessions for condition 1, ref. delete_share_pages() */
|
xizi_task_manager.task_yield_noschedule(cur_cpu()->task, false);
|
||||||
// close all server_sessions
|
|
||||||
struct server_session* server_session = NULL;
|
|
||||||
while (!IS_DOUBLE_LIST_EMPTY(&ptask->svr_sess_listhead)) {
|
|
||||||
server_session = CONTAINER_OF(ptask->svr_sess_listhead.next, struct server_session, node);
|
|
||||||
// cut the connection from task to session
|
|
||||||
if (!server_session->closed) {
|
|
||||||
xizi_share_page_manager.unmap_task_share_pages(ptask, server_session->buf_addr, CLIENT_SESSION_BACKEND(server_session)->nr_pages);
|
|
||||||
server_session->closed = true;
|
|
||||||
}
|
|
||||||
doubleListDel(&server_session->node);
|
|
||||||
SERVER_SESSION_BACKEND(server_session)->server = NULL;
|
|
||||||
// delete session (also cut connection from session to task)
|
|
||||||
if (SERVER_SESSION_BACKEND(server_session)->client_side.closed) {
|
|
||||||
xizi_share_page_manager.delete_share_pages(SERVER_SESSION_BACKEND(server_session));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// close all client_sessions
|
|
||||||
struct client_session* client_session = NULL;
|
|
||||||
while (!IS_DOUBLE_LIST_EMPTY(&ptask->cli_sess_listhead)) {
|
|
||||||
client_session = CONTAINER_OF(ptask->cli_sess_listhead.next, struct client_session, node);
|
|
||||||
// cut the connection from task to session
|
|
||||||
if (!client_session->closed) {
|
|
||||||
xizi_share_page_manager.unmap_task_share_pages(ptask, client_session->buf_addr, CLIENT_SESSION_BACKEND(client_session)->nr_pages);
|
|
||||||
client_session->closed = true;
|
|
||||||
}
|
|
||||||
doubleListDel(&client_session->node);
|
|
||||||
CLIENT_SESSION_BACKEND(client_session)->client = NULL;
|
|
||||||
// delete session (also cut connection from session to task)
|
|
||||||
if (CLIENT_SESSION_BACKEND(client_session)->server_side.closed) {
|
|
||||||
xizi_share_page_manager.delete_share_pages(CLIENT_SESSION_BACKEND(client_session));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (ptask->server_identifier.meta != NULL) {
|
|
||||||
struct TraceTag server_identifier_owner;
|
|
||||||
AchieveResourceTag(&server_identifier_owner, RequireRootTag(), "softkernel/server-identifier");
|
|
||||||
assert(server_identifier_owner.meta != NULL);
|
|
||||||
DeleteResource(&ptask->server_identifier, &server_identifier_owner);
|
|
||||||
}
|
|
||||||
|
|
||||||
// delete task for pcb_list
|
|
||||||
xizi_task_manager.task_yield_noschedule(ptask, true);
|
|
||||||
ptask->state = DEAD;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
|
@ -79,26 +79,33 @@ int user_irq_handler(int irq, void* tf, void* arg)
|
||||||
AchieveResourceTag(&mmu_driver_tag, RequireRootTag(), "/hardkernel/mmu-ac-resource");
|
AchieveResourceTag(&mmu_driver_tag, RequireRootTag(), "/hardkernel/mmu-ac-resource");
|
||||||
p_mmu_driver = (struct MmuCommonDone*)AchieveResource(&mmu_driver_tag);
|
p_mmu_driver = (struct MmuCommonDone*)AchieveResource(&mmu_driver_tag);
|
||||||
}
|
}
|
||||||
p_mmu_driver->LoadPgdir((uintptr_t)V2P(kernel_irq_proxy->pgdir.pd_addr));
|
|
||||||
send_irq_to_user(irq);
|
|
||||||
p_mmu_driver->LoadPgdir((uintptr_t)V2P(cur_cpu()->task->pgdir.pd_addr));
|
|
||||||
|
|
||||||
next_task_emergency = irq_forward_table[irq].handle_task;
|
if (irq_forward_table[irq].handle_task != NULL) {
|
||||||
xizi_task_manager.task_yield_noschedule(cur_cpu()->task, false);
|
p_mmu_driver->LoadPgdir((uintptr_t)V2P(kernel_irq_proxy->pgdir.pd_addr));
|
||||||
|
send_irq_to_user(irq);
|
||||||
|
p_mmu_driver->LoadPgdir((uintptr_t)V2P(cur_cpu()->task->pgdir.pd_addr));
|
||||||
|
|
||||||
|
next_task_emergency = irq_forward_table[irq].handle_task;
|
||||||
|
xizi_task_manager.task_yield_noschedule(cur_cpu()->task, false);
|
||||||
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
extern int create_session_inner(struct TaskMicroDescriptor* client, struct TaskMicroDescriptor* server, int capacity, struct Session* user_session);
|
extern int create_session_inner(struct TaskMicroDescriptor* client, struct TaskMicroDescriptor* server, int capacity, struct Session* user_session);
|
||||||
/// @warning no tested.
|
/// @warning no tested.
|
||||||
|
|
||||||
|
static struct XiziTrapDriver* p_intr_driver = NULL;
|
||||||
int sys_register_irq(int irq_num, int irq_opcode)
|
int sys_register_irq(int irq_num, int irq_opcode)
|
||||||
{
|
{
|
||||||
// init intr resource;
|
// init intr resource;
|
||||||
static struct TraceTag intr_ac_tag;
|
if (p_intr_driver == NULL) {
|
||||||
if (!AchieveResourceTag(&intr_ac_tag, RequireRootTag(), "hardkernel/intr-ac-resource")) {
|
struct TraceTag intr_ac_tag;
|
||||||
ERROR("intr not initialized.\n");
|
if (!AchieveResourceTag(&intr_ac_tag, RequireRootTag(), "hardkernel/intr-ac-resource")) {
|
||||||
return -1;
|
ERROR("intr not initialized.\n");
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
p_intr_driver = (struct XiziTrapDriver*)AchieveResource(&intr_ac_tag);
|
||||||
}
|
}
|
||||||
struct XiziTrapDriver* p_intr_driver = AchieveResource(&intr_ac_tag);
|
|
||||||
|
|
||||||
// init kerenl sender proxy
|
// init kerenl sender proxy
|
||||||
if (kernel_irq_proxy == NULL) {
|
if (kernel_irq_proxy == NULL) {
|
||||||
|
@ -118,6 +125,30 @@ int sys_register_irq(int irq_num, int irq_opcode)
|
||||||
irq_forward_table[irq_num].opcode = irq_opcode;
|
irq_forward_table[irq_num].opcode = irq_opcode;
|
||||||
create_session_inner(kernel_irq_proxy, cur_task, PAGE_SIZE, &irq_forward_table[irq_num].session);
|
create_session_inner(kernel_irq_proxy, cur_task, PAGE_SIZE, &irq_forward_table[irq_num].session);
|
||||||
p_intr_driver->bind_irq_handler(irq_num, user_irq_handler);
|
p_intr_driver->bind_irq_handler(irq_num, user_irq_handler);
|
||||||
|
cur_task->bind_irq = true;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int sys_unbind_irq(struct TaskMicroDescriptor* task, int irq_num)
|
||||||
|
{
|
||||||
|
if (irq_forward_table[irq_num].handle_task != task) {
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
irq_forward_table[irq_num].handle_task = NULL;
|
||||||
|
sys_close_session(&irq_forward_table[irq_num].session);
|
||||||
|
DEBUG("Unbind: %s to irq %d", task->name, irq_num);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int sys_unbind_irq_all(struct TaskMicroDescriptor* task)
|
||||||
|
{
|
||||||
|
for (int idx = 0; idx < NR_IRQS; idx++) {
|
||||||
|
if (irq_forward_table[idx].handle_task == task) {
|
||||||
|
sys_unbind_irq(task, idx);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
task->bind_irq = false;
|
||||||
|
return 0;
|
||||||
|
}
|
|
@ -38,12 +38,10 @@ struct TaskMicroDescriptor* max_priority_runnable_task(void)
|
||||||
|
|
||||||
DOUBLE_LIST_FOR_EACH_ENTRY(task, &xizi_task_manager.task_list_head[priority], node)
|
DOUBLE_LIST_FOR_EACH_ENTRY(task, &xizi_task_manager.task_list_head[priority], node)
|
||||||
{
|
{
|
||||||
if (task->state == READY) {
|
if (task->state == READY && !task->dead) {
|
||||||
// found a runnable task, stop this look up
|
// found a runnable task, stop this look up
|
||||||
return task;
|
return task;
|
||||||
} else if (task->state == DEAD) {
|
} else if (task->dead && task->state != RUNNING) {
|
||||||
// found a killed task, stop this loop
|
|
||||||
// change in pcb_list may break this loop, so find a runnable in next look up
|
|
||||||
xizi_task_manager.free_pcb(task);
|
xizi_task_manager.free_pcb(task);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
@ -57,13 +55,10 @@ struct TaskMicroDescriptor* round_robin_runnable_task(uint32_t priority)
|
||||||
|
|
||||||
DOUBLE_LIST_FOR_EACH_ENTRY(task, &xizi_task_manager.task_list_head[priority], node)
|
DOUBLE_LIST_FOR_EACH_ENTRY(task, &xizi_task_manager.task_list_head[priority], node)
|
||||||
{
|
{
|
||||||
|
if (task->state == READY && !task->dead) {
|
||||||
if (task->state == READY) {
|
|
||||||
// found a runnable task, stop this look up
|
// found a runnable task, stop this look up
|
||||||
return task;
|
return task;
|
||||||
} else if (task->state == DEAD) {
|
} else if (task->dead && task->state != RUNNING) {
|
||||||
// found a killed task, stop this loop
|
|
||||||
// change in pcb_list may break this loop, so find a runnable in next look up
|
|
||||||
xizi_task_manager.free_pcb(task);
|
xizi_task_manager.free_pcb(task);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
|
@ -32,10 +32,11 @@ Modification:
|
||||||
#include "core.h"
|
#include "core.h"
|
||||||
|
|
||||||
#include "assert.h"
|
#include "assert.h"
|
||||||
|
#include "kalloc.h"
|
||||||
#include "log.h"
|
#include "log.h"
|
||||||
#include "multicores.h"
|
#include "multicores.h"
|
||||||
#include "kalloc.h"
|
|
||||||
#include "scheduler.h"
|
#include "scheduler.h"
|
||||||
|
#include "syscall.h"
|
||||||
#include "task.h"
|
#include "task.h"
|
||||||
|
|
||||||
struct CPU global_cpus[NR_CPU];
|
struct CPU global_cpus[NR_CPU];
|
||||||
|
@ -74,6 +75,59 @@ static struct TaskMicroDescriptor* _alloc_task_cb()
|
||||||
return task;
|
return task;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int _task_retrieve_sys_resources(struct TaskMicroDescriptor* ptask)
|
||||||
|
{
|
||||||
|
assert(ptask != NULL);
|
||||||
|
|
||||||
|
/* handle sessions for condition 1, ref. delete_share_pages() */
|
||||||
|
// close all server_sessions
|
||||||
|
struct server_session* server_session = NULL;
|
||||||
|
while (!IS_DOUBLE_LIST_EMPTY(&ptask->svr_sess_listhead)) {
|
||||||
|
server_session = CONTAINER_OF(ptask->svr_sess_listhead.next, struct server_session, node);
|
||||||
|
// cut the connection from task to session
|
||||||
|
if (!server_session->closed) {
|
||||||
|
xizi_share_page_manager.unmap_task_share_pages(ptask, server_session->buf_addr, CLIENT_SESSION_BACKEND(server_session)->nr_pages);
|
||||||
|
server_session->closed = true;
|
||||||
|
}
|
||||||
|
doubleListDel(&server_session->node);
|
||||||
|
SERVER_SESSION_BACKEND(server_session)->server = NULL;
|
||||||
|
// delete session (also cut connection from session to task)
|
||||||
|
if (SERVER_SESSION_BACKEND(server_session)->client_side.closed) {
|
||||||
|
xizi_share_page_manager.delete_share_pages(SERVER_SESSION_BACKEND(server_session));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// close all client_sessions
|
||||||
|
struct client_session* client_session = NULL;
|
||||||
|
while (!IS_DOUBLE_LIST_EMPTY(&ptask->cli_sess_listhead)) {
|
||||||
|
client_session = CONTAINER_OF(ptask->cli_sess_listhead.next, struct client_session, node);
|
||||||
|
// cut the connection from task to session
|
||||||
|
if (!client_session->closed) {
|
||||||
|
xizi_share_page_manager.unmap_task_share_pages(ptask, client_session->buf_addr, CLIENT_SESSION_BACKEND(client_session)->nr_pages);
|
||||||
|
client_session->closed = true;
|
||||||
|
}
|
||||||
|
doubleListDel(&client_session->node);
|
||||||
|
CLIENT_SESSION_BACKEND(client_session)->client = NULL;
|
||||||
|
// delete session (also cut connection from session to task)
|
||||||
|
if (CLIENT_SESSION_BACKEND(client_session)->server_side.closed) {
|
||||||
|
xizi_share_page_manager.delete_share_pages(CLIENT_SESSION_BACKEND(client_session));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ptask->server_identifier.meta != NULL) {
|
||||||
|
struct TraceTag server_identifier_owner;
|
||||||
|
AchieveResourceTag(&server_identifier_owner, RequireRootTag(), "softkernel/server-identifier");
|
||||||
|
assert(server_identifier_owner.meta != NULL);
|
||||||
|
DeleteResource(&ptask->server_identifier, &server_identifier_owner);
|
||||||
|
}
|
||||||
|
|
||||||
|
// delete registered irq if there is one
|
||||||
|
if (ptask->bind_irq) {
|
||||||
|
sys_unbind_irq_all(ptask);
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
/// @brief this function changes task list without locking, so it must be called inside a lock critical area
|
/// @brief this function changes task list without locking, so it must be called inside a lock critical area
|
||||||
/// @param task
|
/// @param task
|
||||||
static void _dealloc_task_cb(struct TaskMicroDescriptor* task)
|
static void _dealloc_task_cb(struct TaskMicroDescriptor* task)
|
||||||
|
@ -83,6 +137,8 @@ static void _dealloc_task_cb(struct TaskMicroDescriptor* task)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
_task_retrieve_sys_resources(task);
|
||||||
|
|
||||||
// stack is mapped in vspace, so it should be free by pgdir
|
// stack is mapped in vspace, so it should be free by pgdir
|
||||||
if (task->pgdir.pd_addr) {
|
if (task->pgdir.pd_addr) {
|
||||||
xizi_pager.free_user_pgdir(&task->pgdir);
|
xizi_pager.free_user_pgdir(&task->pgdir);
|
||||||
|
@ -112,7 +168,7 @@ static void _dealloc_task_cb(struct TaskMicroDescriptor* task)
|
||||||
extern void trap_return(void);
|
extern void trap_return(void);
|
||||||
void task_prepare_enter()
|
void task_prepare_enter()
|
||||||
{
|
{
|
||||||
spinlock_unlock(&whole_kernel_lock);
|
xizi_leave_kernel();
|
||||||
trap_return();
|
trap_return();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -124,10 +180,7 @@ static struct TaskMicroDescriptor* _new_task_cb()
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
// init vm
|
// init vm
|
||||||
if (!xizi_pager.new_pgdir(&task->pgdir)) {
|
task->pgdir.pd_addr = NULL;
|
||||||
_dealloc_task_cb(task);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
/* init basic task member */
|
/* init basic task member */
|
||||||
doubleListNodeInit(&task->cli_sess_listhead);
|
doubleListNodeInit(&task->cli_sess_listhead);
|
||||||
doubleListNodeInit(&task->svr_sess_listhead);
|
doubleListNodeInit(&task->svr_sess_listhead);
|
||||||
|
@ -180,14 +233,12 @@ static void _scheduler(struct SchedulerRightGroup right_group)
|
||||||
assert(cur_cpu()->task == NULL);
|
assert(cur_cpu()->task == NULL);
|
||||||
if (next_task_emergency != NULL && next_task->state == READY) {
|
if (next_task_emergency != NULL && next_task->state == READY) {
|
||||||
next_task = next_task_emergency;
|
next_task = next_task_emergency;
|
||||||
next_task->state = RUNNING;
|
|
||||||
} else {
|
} else {
|
||||||
next_task = xizi_task_manager.next_runnable_task();
|
next_task = xizi_task_manager.next_runnable_task();
|
||||||
}
|
}
|
||||||
next_task_emergency = NULL;
|
next_task_emergency = NULL;
|
||||||
if (next_task != NULL) {
|
if (next_task != NULL) {
|
||||||
assert(next_task->state == READY);
|
assert(next_task->state == READY);
|
||||||
next_task->state = RUNNING;
|
|
||||||
}
|
}
|
||||||
spinlock_unlock(&whole_kernel_lock);
|
spinlock_unlock(&whole_kernel_lock);
|
||||||
|
|
||||||
|
@ -199,11 +250,16 @@ static void _scheduler(struct SchedulerRightGroup right_group)
|
||||||
|
|
||||||
/* a runnable task */
|
/* a runnable task */
|
||||||
spinlock_lock(&whole_kernel_lock);
|
spinlock_lock(&whole_kernel_lock);
|
||||||
assert(next_task->state == RUNNING);
|
if (next_task->state == READY) {
|
||||||
|
next_task->state = RUNNING;
|
||||||
|
} else {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
struct CPU* cpu = cur_cpu();
|
struct CPU* cpu = cur_cpu();
|
||||||
cpu->task = next_task;
|
cpu->task = next_task;
|
||||||
p_mmu_driver->LoadPgdir((uintptr_t)V2P(next_task->pgdir.pd_addr));
|
p_mmu_driver->LoadPgdir((uintptr_t)V2P(next_task->pgdir.pd_addr));
|
||||||
context_switch(&cpu->scheduler, next_task->main_thread.context);
|
context_switch(&cpu->scheduler, next_task->main_thread.context);
|
||||||
|
assert(next_task->state != RUNNING);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -62,8 +62,7 @@ void intr_irq_dispatch(struct trapframe* tf)
|
||||||
assert(p_intr_driver != NULL);
|
assert(p_intr_driver != NULL);
|
||||||
uintptr_t int_info = 0;
|
uintptr_t int_info = 0;
|
||||||
if ((int_info = p_intr_driver->hw_before_irq()) == 0) {
|
if ((int_info = p_intr_driver->hw_before_irq()) == 0) {
|
||||||
xizi_leave_kernel();
|
goto intr_leave_interrupt;
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
struct TaskMicroDescriptor* current_task = cur_cpu()->task;
|
struct TaskMicroDescriptor* current_task = cur_cpu()->task;
|
||||||
|
@ -77,7 +76,7 @@ void intr_irq_dispatch(struct trapframe* tf)
|
||||||
|
|
||||||
// distribute irq
|
// distribute irq
|
||||||
irq_handler_t isr = p_intr_driver->sw_irqtbl[irq].handler;
|
irq_handler_t isr = p_intr_driver->sw_irqtbl[irq].handler;
|
||||||
if (isr) {
|
if (isr != NULL) {
|
||||||
isr(irq, tf, NULL);
|
isr(irq, tf, NULL);
|
||||||
} else {
|
} else {
|
||||||
default_interrupt_routine();
|
default_interrupt_routine();
|
||||||
|
@ -93,6 +92,7 @@ void intr_irq_dispatch(struct trapframe* tf)
|
||||||
}
|
}
|
||||||
assert(current_task == cur_cpu()->task);
|
assert(current_task == cur_cpu()->task);
|
||||||
|
|
||||||
|
intr_leave_interrupt:
|
||||||
xizi_leave_kernel();
|
xizi_leave_kernel();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -107,3 +107,8 @@ void xizi_leave_kernel()
|
||||||
spinlock_unlock(&whole_kernel_lock);
|
spinlock_unlock(&whole_kernel_lock);
|
||||||
p_intr_driver->cpu_irq_enable();
|
p_intr_driver->cpu_irq_enable();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool xizi_is_in_kernel()
|
||||||
|
{
|
||||||
|
return is_spinlock_locked(&whole_kernel_lock);
|
||||||
|
}
|
Loading…
Reference in New Issue