Chinaunix首页 | 论坛 | 博客
  • 博客访问: 15308429
  • 博文数量: 2005
  • 博客积分: 11986
  • 博客等级: 上将
  • 技术积分: 22535
  • 用 户 组: 普通用户
  • 注册时间: 2007-05-17 13:56
文章分类

全部博文(2005)

文章存档

2014年(2)

2013年(2)

2012年(16)

2011年(66)

2010年(368)

2009年(743)

2008年(491)

2007年(317)

分类: LINUX

2008-11-06 14:16:12

浅析linux内核中timer定时器的生成和sofirq软中断调用流程

  mod_timer添加的定时器timer在内核的软中断中发生调用,__run_timers会spin_lock_irq(&base->lock);禁止cpu中断,所以我们的timer回调处理函数handler工作在irq关闭的环境中,所以需要作很多考虑,比如在handler中尽量不要执行会引起pending的函数调用,比如kmalloc之类可能引起pending的操作,否则会使kernel永远停在我们的handler中不能返回,这样kernel将因为我们ko设计上的失败而死机[luther.gliethttp]!
  我们可以使用如下几行语句,向我们的ko驱动添加一个timer定时器,来处理时间事件:
struct __wlanwlan_check_tx_flow_timer
{
    struct timer_list timer;
    int timer_freq;
} wlan_check_tx_flow_timer = {
        .timer_freq = 8*1000,
};
static void wlan_check_tx_flow_timer_handler(unsigned long data)
{
    ...
    //重新启动timer定时器
    mod_timer(&wlan_check_tx_flow_timer.timer, jiffies + msecs_to_jiffies(wlan_check_tx_flow_timer.timer_freq));
    ...
}
//设置定时器
setup_timer(&wlan_check_tx_flow_timer.timer, wlan_check_tx_flow_timer_handler, (unsigned long)&wlan_check_tx_flow_timer);
//添加定时器
mod_timer(&wlan_check_tx_flow_timer.timer, jiffies + msecs_to_jiffies(wlan_check_tx_flow_timer.timer_freq));

那么这个wlan_check_tx_flow_timer_handler处理函数在什么时候被调用的呢?那么我们追入内核中,看看kernel对定时器的具体管理.
首先kernel在启动的最前面注册TIMER_SOFTIRQ的处理函数[luther.gliethttp],
start_kernel
=>init_timers
=>open_softirq(TIMER_SOFTIRQ, run_timer_softirq, NULL);
那么由谁来调用raise_softirq(TIMER_SOFTIRQ);触发TIMER_SOFTIRQ软中断呢,这就和平台相关了,对于pxa935处理器来说[luther.gliethttp],
MACHINE_START(LUTHER, "luther")
    .phys_io = 0x40000000,
    .boot_params = 0xa0000100,
    .io_pg_offst = (io_p2v(0x40000000) >> 18) & 0xfffc,
    .map_io = pxa_map_io,
    .init_irq = pxa3xx_init_irq,
    .timer = &pxa_timer,
    .init_machine = luther_init,
MACHINE_END
=>pxa_timer_init//平台对应的定时器初始化
==>pxa_timer_irq.dev_id = &ckevt_32ktimer;
==>setup_irq(IRQ_OST_4_11, &pxa_timer_irq); //32768的rtc
==>clockevents_register_device(&ckevt_32ktimer);

pxa_timer_interrupt中断处理函数
=>c->event_handler(c);也就是tick_handle_periodic系统时钟函数
=>tick_handle_periodic
=>update_process_times
=>run_local_timers
=>raise_softirq(TIMER_SOFTIRQ);
这里仅仅是触发了TIMER_SOFTIRQ软中断,那么在什么地方处理我们mod_timer添加的timer定时器处理函数wlan_check_tx_flow_timer_handler呢[luther.gliethttp]
__irq_svc://内核中发生的中断
__irq_usr://用户空间时发生的中断
=>asm_do_IRQ
=>irq_exit
=>do_softirq
=>__do_softirq
=>调用上面注册的run_timer_softirq软中断处理函数
=>run_timer_softirq
=>__run_timers
static inline void __run_timers(struct tvec_base *base)
{
    struct timer_list *timer;

    spin_lock_irq(&base->lock);//禁止中断
    while (time_after_eq(jiffies, base->timer_jiffies)) {
        ...
        if (时间到了) {
        ...
        fn = timer->function;
        data = timer->data;
        fn(data);//这就是我们上面添加的static void wlan_check_tx_flow_timer_handler(unsigned long data);定时器处理函数了.
        ...
        }
        ...
    }
    set_running_timer(base, NULL);
    spin_unlock_irq(&base->lock);//打开中断
}
//================
include/asm/hardirq.h
typedef struct {
    unsigned int __softirq_pending;
    unsigned int local_timer_irqs;
} ____cacheline_aligned irq_cpustat_t;
//================
kernel/softirq.c|45| irq_cpustat_t irq_stat[NR_CPUS] ____cacheline_aligned;
#ifndef __ARCH_IRQ_STAT
irq_cpustat_t irq_stat[NR_CPUS] ____cacheline_aligned;//在这里定义irq_stat存储空间
EXPORT_SYMBOL(irq_stat);
#endif
//================
include/linux/irq_cpustat.h
#ifndef __ARCH_IRQ_STAT
//引用的就是上面的irq_cpustat_t irq_stat[NR_CPUS] ____cacheline_aligned;
extern irq_cpustat_t irq_stat[];        /* defined in asm/hardirq.h */
#define __IRQ_STAT(cpu, member)    (irq_stat[cpu].member)
#endif
//================
arch/arm/kernel/entry-armv.S|331| .word    irq_stat

#ifdef CONFIG_PREEMPT
svc_preempt:
    teq    r8, #0                @ was preempt count = 0
    ldreq    r6, .LCirq_stat //操作
    movne    pc, lr                @ no
    ldr    r0, [r6, #4]            @ local_irq_count
    ldr    r1, [r6, #8]            @ local_bh_count
    adds    r0, r0, r1
    movne    pc, lr
    mov    r7, #0                @ preempt_schedule_irq
    str    r7, [tsk, #TI_PREEMPT]        @ expects preempt_count == 0
1:    bl    preempt_schedule_irq        @ irq en/disable is done inside
    ldr    r0, [tsk, #TI_FLAGS]        @ get new tasks TI_FLAGS
    tst    r0, #_TIF_NEED_RESCHED
    beq    preempt_return            @ go again
    b    1b
#endif

.LCirq_stat:
    .word    irq_stat //引用irq_cpustat_t irq_stat[NR_CPUS] ____cacheline_aligned;地址
#endif

  /* arch independent irq_stat fields */
#define local_softirq_pending() \
    __IRQ_STAT(smp_processor_id(), __softirq_pending)

#define __ARCH_IRQ_EXIT_IRQS_DISABLED    1
#ifdef __ARCH_IRQ_EXIT_IRQS_DISABLED
# define invoke_softirq()    __do_softirq() //是这个
#else
# define invoke_softirq()    do_softirq()
#endif

#ifndef __ARCH_SET_SOFTIRQ_PENDING
#define set_softirq_pending(x) (local_softirq_pending() = (x))
#define or_softirq_pending(x) (local_softirq_pending() |= (x))
#endif
#define __raise_softirq_irqoff(nr) do { or_softirq_pending(1UL << (nr)); } while (0)

inline void raise_softirq_irqoff(unsigned int nr)
{
    __raise_softirq_irqoff(nr);
        if (!in_interrupt())
        wakeup_softirqd();
}
void raise_softirq(unsigned int nr)
{
    unsigned long flags;

    local_irq_save(flags);
    raise_softirq_irqoff(nr);
    local_irq_restore(flags);
}


=>s3c2410_timer_interrupt
=>timer_tick

=>pxa_timer_init
==>pxa_timer_irq.dev_id = &ckevt_32ktimer;
==>setup_irq(IRQ_OST_4_11, &pxa_timer_irq); //32768的rtc
==>clockevents_register_device(&ckevt_32ktimer);
=>clockevents_register_device
=>clockevents_do_notify
=>raw_notifier_call_chain(&clockevents_chain, reason, dev);
=>__raw_notifier_call_chain
=>notifier_call_chain(&nh->head, val, v, nr_to_call, nr_calls);
=>nb->notifier_call(nb, val, v);就是tick_notify

start_kernel
=>tick_init

static struct notifier_block tick_notifier = {
    .notifier_call = tick_notify,
};
void __init tick_init(void)
{
    clockevents_register_notifier(&tick_notifier);
}
clockevents_register_notifier
=>raw_notifier_chain_register(&clockevents_chain, nb);
=>notifier_chain_register将tick_notifier添加到clockevents_chain这个单向链表中[luther.gliethttp]
static int tick_notify(struct notifier_block *nb, unsigned long reason,
             void *dev)
{
    switch (reason) {

    case CLOCK_EVT_NOTIFY_ADD:
        return tick_check_new_device(dev);
    ...
    return NOTIFY_OK;
}

=>tick_notify
=>tick_check_new_device
=>tick_setup_device(td, newdev, cpu, cpumask);
static void tick_setup_device(struct tick_device *td,
             struct clock_event_device *newdev, int cpu,
             cpumask_t cpumask)
{
    ktime_t next_event;
    void (*handler)(struct clock_event_device *) = NULL;

    /*
     * First device setup ?
     */

    if (!td->evtdev) {
        /*
         * If no cpu took the do_timer update, assign it to
         * this cpu:
         */

        if (tick_do_timer_cpu == -1) {
            tick_do_timer_cpu = cpu;
            tick_next_period = ktime_get();
            tick_period = ktime_set(0, NSEC_PER_SEC / HZ);
        }

        /*
         * Startup in periodic mode first.
         */

        td->mode = TICKDEV_MODE_PERIODIC;//设置第1个tick设备为TICKDEV_MODE_PERIODIC模式
    } else {
        handler = td->evtdev->event_handler;
        next_event = td->evtdev->next_event;
    }

    td->evtdev = newdev;
    ...
    if (td->mode == TICKDEV_MODE_PERIODIC)
        tick_setup_periodic(newdev, 0);
    else
        tick_setup_oneshot(newdev, handler, next_event);
}

void tick_setup_periodic(struct clock_event_device *dev, int broadcast)
{
    tick_set_periodic_handler(dev, broadcast);//设置event_handler处理函数为dev->event_handler = tick_handle_periodic;

    /* Broadcast setup ? */
    if (!tick_device_is_functional(dev))
        return;

    if (dev->features & CLOCK_EVT_FEAT_PERIODIC) {
        clockevents_set_mode(dev, CLOCK_EVT_MODE_PERIODIC);
    } else {
        unsigned long seq;
        ktime_t next;

        do {
            seq = read_seqbegin(&xtime_lock);
            next = tick_next_period;
        } while (read_seqretry(&xtime_lock, seq));

        clockevents_set_mode(dev, CLOCK_EVT_MODE_ONESHOT);

        for (;;) {
            if (!clockevents_program_event(dev, next, ktime_get()))
                return;
            next = ktime_add(next, tick_period);
        }
    }
}

void tick_set_periodic_handler(struct clock_event_device *dev, int broadcast)
{
    if (!broadcast)
        dev->event_handler = tick_handle_periodic;
    else
        dev->event_handler = tick_handle_periodic_broadcast;
}

=>pxa_timer_interrupt
{
    ...
    if (OSSR & OST_C4) {
        OIER &= ~OST_C4;
        OSSR = OST_C4;
        if (timer32k_enabled)
            c->event_handler(c);//调用tick_handle_periodic处理函数,作为
    }
    ...
}

void tick_handle_periodic(struct clock_event_device *dev)
{
    int cpu = smp_processor_id();
    ktime_t next;

    tick_periodic(cpu);//调用do_timer(1);将jiffies_64加1

    if (dev->mode != CLOCK_EVT_MODE_ONESHOT)
        return;
    /*
     * Setup the next period for devices, which do not have
     * periodic mode:
     */

    next = ktime_add(dev->next_event, tick_period);
    for (;;) {
        if (!clockevents_program_event(dev, next, ktime_get()))
            return;
        tick_periodic(cpu);
        next = ktime_add(next, tick_period);
    }
}

static void tick_periodic(int cpu)
{
    if (tick_do_timer_cpu == cpu) {
        write_seqlock(&xtime_lock);

        /* Keep track of the next tick event */
        tick_next_period = ktime_add(tick_next_period, tick_period);

        do_timer(1);
        write_sequnlock(&xtime_lock);
    }

    update_process_times(user_mode(get_irq_regs()));
    profile_tick(CPU_PROFILING);
}

arch/arm/kernel/time.c|332| update_process_times(user_mode(get_irq_regs()));
=>update_process_times
=>run_local_timers
=>raise_softirq(TIMER_SOFTIRQ);//触发软中断,当irq_exit时调用__do_softirq来处理


=>run_timer_softirq
=>__run_timers
=>
fn = timer->function;//执行
data = timer->data;
fn(data);
//================
include/asm/arch-pxa/entry-macro.S|22| .macro    get_irqnr_and_base, irqnr, irqstat, base, tmp
//pxa获取irq中断号函数
//================
arch/arm/kernel/entry-armv.S|37| bne    asm_do_IRQ
    .macro    irq_handler
    get_irqnr_preamble r5, lr
1:    get_irqnr_and_base r0, r6, r5, lr //获取irq中断号,存储到r0寄存器中,作为参数传递给asm_do_IRQ
    movne    r1, sp
    @
    @ routine called with r0 = irq number, r1 = struct pt_regs *
    @
    adrne    lr, 1b
    bne    asm_do_IRQ
    ...
//================
    .align    5
__irq_svc://内核中发生的中断
    svc_entry
    ...
    irq_handler
    ...
//================
    .align    5
__irq_usr://用户空间时发生的中断
    usr_entry
    ...
    irq_handler
    ...
//================
    .macro    vector_stub, name, mode, correction=0
    .align    5

vector_\name:
    .if \correction
    sub    lr, lr, #\correction
    .endif

    @
    @ Save r0, lr_<exception> (parent PC) and spsr_<exception>
    @ (parent CPSR)
    @
    stmia    sp, {r0, lr}        @ save r0, lr
    mrs    lr, spsr
    str    lr, [sp, #8]        @ save spsr

    @
    @ Prepare for SVC32 mode. IRQs remain disabled.
    @
    mrs    r0, cpsr
    eor    r0, r0, #(\mode ^ SVC_MODE)
    msr    spsr_cxsf, r0

    @
    @ the branch table must immediately follow this code
    @
    and    lr, lr, #0x0f //lr存储了spsr,所以一共有16种cpu模式
    mov    r0, sp //传参
    ldr    lr, [pc, lr, lsl #2]//取出相应模式下的处理函数指针,比如__irq_usr或者__irq_svc
    movs    pc, lr            @ branch to handler in SVC mode
    .endm
//================
    .globl    __stubs_start
__stubs_start:
/*
 * Interrupt dispatcher
 */

    vector_stub    irq, IRQ_MODE, 4

    .long    __irq_usr            @ 0 (USR_26 / USR_32)
    .long    __irq_invalid        @ 1 (FIQ_26 / FIQ_32)
    .long    __irq_invalid        @ 2 (IRQ_26 / IRQ_32)
    .long    __irq_svc            @ 3 (SVC_26 / SVC_32)
    .long    __irq_invalid            @ 4
    .long    __irq_invalid            @ 5
    .long    __irq_invalid            @ 6
    .long    __irq_invalid            @ 7
    .long    __irq_invalid            @ 8
    .long    __irq_invalid            @ 9
    .long    __irq_invalid            @ a
    .long    __irq_invalid            @ b
    .long    __irq_invalid            @ c
    .long    __irq_invalid            @ d
    .long    __irq_invalid            @ e
    .long    __irq_invalid            @ f
//================
    .globl    __vectors_start
__vectors_start:
    swi    SYS_ERROR0
    b    vector_und + stubs_offset
    ldr    pc, .LCvswi + stubs_offset
    b    vector_pabt + stubs_offset
    b    vector_dabt + stubs_offset
    b    vector_addrexcptn + stubs_offset
    b    vector_irq + stubs_offset
    b    vector_fiq + stubs_offset
//================
asm_do_IRQ(unsigned int irq, struct pt_regs *regs)
=>desc_handle_irq(irq, desc);//
static inline void desc_handle_irq(unsigned int irq, struct irq_desc *desc)
{
    desc->handle_irq(irq, desc);//调用中断号irq对应的handler回调处理函数[luther.gliethttp]
}
__irq_svc://内核中发生的中断
__irq_usr://用户空间时发生的中断
=>asm_do_IRQ
=>irq_exit
=>do_softirq
=>__do_softirq
=>
{
    ...
    h = softirq_vec;//执行软中断函数

    do {
        if (pending & 1) {
            h->action(h);
//如果32768的时间到达,那asm_do_IRQ中将触发raise_softirq(TIMER_SOFTIRQ);
//在这里将执行管理系统tick的run_timer_softirq软中断[luther.gliethttp]
            rcu_bh_qsctr_inc(cpu);
        }
        h++;
        pending >>= 1;
    } while (pending);
    ...
}

start_kernel
=>init_timers
=>open_softirq(TIMER_SOFTIRQ, run_timer_softirq, NULL);
void open_softirq(int nr, void (*action)(struct softirq_action*), void *data)
{
    softirq_vec[nr].data = data;
    softirq_vec[nr].action = action;
}

static void run_timer_softirq(struct softirq_action *h)
{
    struct tvec_base *base = __get_cpu_var(tvec_bases);//获得time时间根

    hrtimer_run_pending();

    if (time_after_eq(jiffies, base->timer_jiffies))
        __run_timers(base);
}
//执行软中断
=>run_timer_softirq
=>__run_timers
=>
fn = timer->function;
data = timer->data;
fn(data);

static inline void __run_timers(struct tvec_base *base)
{
    ...
    spin_lock_irq(&base->lock);//禁止中断
    ...
    fn = timer->function;
    data = timer->data;
    fn(data);
    ...
    set_running_timer(base, NULL);
    spin_unlock_irq(&base->lock);//打开中断
}

mod_timer
=>__mod_timer
int __mod_timer(struct timer_list *timer, unsigned long expires)
{
    struct tvec_base *base, *new_base;
    unsigned long flags;
    int ret = 0;

    timer_stats_timer_set_start_info(timer);
    BUG_ON(!timer->function);

    base = lock_timer_base(timer, &flags);

    if (timer_pending(timer)) {
        detach_timer(timer, 0);
        ret = 1;
    }

    new_base = __get_cpu_var(tvec_bases);//获得time时间根

    if (base != new_base) {
        /*
         * We are trying to schedule the timer on the local CPU.
         * However we can't change timer's base while it is running,
         * otherwise del_timer_sync() can't detect that the timer's
         * handler yet has not finished. This also guarantees that
         * the timer is serialized wrt itself.
         */

        if (likely(base->running_timer != timer)) {
            /* See the comment in lock_timer_base() */
            timer_set_base(timer, NULL);
            spin_unlock(&base->lock);
            base = new_base;
            spin_lock(&base->lock);
            timer_set_base(timer, base);
        }
    }

    timer->expires = expires;
    internal_add_timer(base, timer);
//添加到链表上,这样当timer超时到达时,run_timer_softirq=>__run_timers软中断中将会回调该处理函数[luther.gliethttp].
    spin_unlock_irqrestore(&base->lock, flags);

    return ret;
}

阅读(11792) | 评论(2) | 转发(2) |
给主人留下些什么吧!~~

gliethttp2010-02-01 15:07:15

谢谢,确实是你说的这样子,__run_timers在进入的时候spin_lock_irq(&base->lock);在执行fn定时器回调函数时,又执行了pin_unlock_irq(&base->lock);,当fn定时器回调函数执行完毕之后,while循环又再次调用spin_lock_irq(&base->lock);禁止了中断,你看得很细,不错!

wilbur5122010-01-31 21:40:57

定时器回调函数的执行环境其实是开启中断,解锁以及抢占的 请楼主查证