Chinaunix首页 | 论坛 | 博客
  • 博客访问: 1056515
  • 博文数量: 573
  • 博客积分: 0
  • 博客等级: 民兵
  • 技术积分: 66
  • 用 户 组: 普通用户
  • 注册时间: 2016-06-28 16:21
文章分类

全部博文(573)

文章存档

2018年(3)

2016年(48)

2015年(522)

分类: LINUX

2015-12-04 09:37:08

alloc_page 代码
alloc_pages()->alloc_pages_node()->__alloc_pages()->__alloc_pages_nodemask()->get_page_from_freelist()->buffered_rmqueue()->__rmqueue()->__rmqueue_smallest()->expand()

include\linux\Gpf.h

#define alloc_pages(gfp_mask, order) \
  alloc_pages_node(numa_node_id(), gfp_mask, order)

 

static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask,
      unsigned int order)
{
 /* Unknown node is current node */
 if (nid < 0)
  nid = numa_node_id();

 return __alloc_pages(gfp_mask, order, node_zonelist(nid, gfp_mask));
}

 

static inline struct page *
__alloc_pages(gfp_t gfp_mask, unsigned int order,
  struct zonelist *zonelist)
{
 return __alloc_pages_nodemask(gfp_mask, order, zonelist, NULL);
}

 

mm/page_alloc.c

struct page *
__alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order,
   struct zonelist *zonelist, nodemask_t *nodemask)
{
 enum zone_type high_zoneidx = gfp_zone(gfp_mask);
 struct zone *preferred_zone;
 struct page *page;
 int migratetype = allocflags_to_migratetype(gfp_mask);

 gfp_mask &= gfp_allowed_mask;

 lockdep_trace_alloc(gfp_mask);

 might_sleep_if(gfp_mask & __GFP_WAIT);

 if (should_fail_alloc_page(gfp_mask, order))
  return NULL;

 /*
  * Check the zones suitable for the gfp_mask contain at least one
  * valid zone. It's possible to have an empty zonelist as a result
  * of GFP_THISNODE and a memoryless node
  */
 if (unlikely(!zonelist->_zonerefs->zone))
  return NULL;

 get_mems_allowed();
 /* The preferred zone is used for statistics later */
 first_zones_zonelist(zonelist, high_zoneidx,
    nodemask ? : &cpuset_current_mems_allowed,
    &preferred_zone);
 if (!preferred_zone) {
  put_mems_allowed();
  return NULL;
 }

 /* First allocation attempt */
 page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, order,
   zonelist, high_zoneidx, ALLOC_WMARK_LOW|ALLOC_CPUSET,
   preferred_zone, migratetype);
 if (unlikely(!page))
  page = __alloc_pages_slowpath(gfp_mask, order,
    zonelist, high_zoneidx, nodemask,
    preferred_zone, migratetype);
 put_mems_allowed();

 trace_mm_page_alloc(page, order, gfp_mask, migratetype);
 return page;
}

 

static struct page *
get_page_from_freelist(gfp_t gfp_mask, nodemask_t *nodemask, unsigned int order,
  struct zonelist *zonelist, int high_zoneidx, int alloc_flags,
  struct zone *preferred_zone, int migratetype)
{
 struct zoneref *z;
 struct page *page = NULL;
 int classzone_idx;
 struct zone *zone;
 nodemask_t *allowednodes = NULL;/* zonelist_cache approximation */
 int zlc_active = 0;  /* set if using zonelist_cache */
 int did_zlc_setup = 0;  /* just call zlc_setup() one time */

 classzone_idx = zone_idx(preferred_zone);
zonelist_scan:
 /*
  * Scan zonelist, looking for a zone with enough free.
  * See also cpuset_zone_allowed() comment in kernel/cpuset.c.
  */
 for_each_zone_zonelist_nodemask(zone, z, zonelist,
      high_zoneidx, nodemask) {
  if (NUMA_BUILD && zlc_active &&
   !zlc_zone_worth_trying(zonelist, z, allowednodes))
    continue;
  if ((alloc_flags & ALLOC_CPUSET) &&
   !cpuset_zone_allowed_softwall(zone, gfp_mask))
    continue;

  BUILD_BUG_ON(ALLOC_NO_WATERMARKS < NR_WMARK);
  if (!(alloc_flags & ALLOC_NO_WATERMARKS)) {
   unsigned long mark;
   int ret;

   mark = zone->watermark[alloc_flags & ALLOC_WMARK_MASK];
   if (zone_watermark_ok(zone, order, mark,
        classzone_idx, alloc_flags))
    goto try_this_zone;

   if (NUMA_BUILD && !did_zlc_setup && nr_online_nodes > 1) {
    /*
     * we do zlc_setup if there are multiple nodes
     * and before considering the first zone allowed
     * by the cpuset.
     */
    allowednodes = zlc_setup(zonelist, alloc_flags);
    zlc_active = 1;
    did_zlc_setup = 1;
   }

   if (zone_reclaim_mode == 0)
    goto this_zone_full;

   /*
    * As we may have just activated ZLC, check if the first
    * eligible zone has failed zone_reclaim recently.
    */
   if (NUMA_BUILD && zlc_active &&
    !zlc_zone_worth_trying(zonelist, z, allowednodes))
    continue;

   ret = zone_reclaim(zone, gfp_mask, order);
   switch (ret) {
   case ZONE_RECLAIM_NOSCAN:
    /* did not scan */
    continue;
   case ZONE_RECLAIM_FULL:
    /* scanned but unreclaimable */
    continue;
   default:
    /* did we reclaim enough */
    if (!zone_watermark_ok(zone, order, mark,
      classzone_idx, alloc_flags))
     goto this_zone_full;
   }
  }

try_this_zone:
  page = buffered_rmqueue(preferred_zone, zone, order,
      gfp_mask, migratetype);
  if (page)
   break;
this_zone_full:
  if (NUMA_BUILD)
   zlc_mark_zone_full(zonelist, z);
 }

 if (unlikely(NUMA_BUILD && page == NULL && zlc_active)) {
  /* Disable zlc cache for second zonelist scan */
  zlc_active = 0;
  goto zonelist_scan;
 }
 return page;
}

 

static inline
struct page *buffered_rmqueue(struct zone *preferred_zone,
   struct zone *zone, int order, gfp_t gfp_flags,
   int migratetype)
{
 unsigned long flags;
 struct page *page;
 int cold = !!(gfp_flags & __GFP_COLD);

again:
 if (likely(order == 0)) {
  struct per_cpu_pages *pcp;
  struct list_head *list;

  local_irq_save(flags);
  pcp = &this_cpu_ptr(zone->pageset)->pcp;
  list = &pcp->lists[migratetype];
  if (list_empty(list)) {
   pcp->count += rmqueue_bulk(zone, 0,
     pcp->batch, list,
     migratetype, cold);
   if (unlikely(list_empty(list)))
    goto failed;
  }

  if (cold)
   page = list_entry(list->prev, struct page, lru);
  else
   page = list_entry(list->next, struct page, lru);

  list_del(&page->lru);
  pcp->count--;
 } else {
  if (unlikely(gfp_flags & __GFP_NOFAIL)) {
   /*
    * __GFP_NOFAIL is not to be used in new code.
    *
    * All __GFP_NOFAIL callers should be fixed so that they
    * properly detect and handle allocation failures.
    *
    * We most definitely don't want callers attempting to
    * allocate greater than order-1 page units with
    * __GFP_NOFAIL.
    */
   WARN_ON_ONCE(order > 1);
  }
  spin_lock_irqsave(&zone->lock, flags);
  page = __rmqueue(zone, order, migratetype);
  spin_unlock(&zone->lock);
  if (!page)
   goto failed;
  __mod_zone_page_state(zone, NR_FREE_PAGES, -(1 << order));
 }

 __count_zone_vm_events(PGALLOC, zone, 1 << order);
 zone_statistics(preferred_zone, zone, gfp_flags);
 local_irq_restore(flags);

 VM_BUG_ON(bad_range(zone, page));
 if (prep_new_page(page, order, gfp_flags))
  goto again;
 return page;

failed:
 local_irq_restore(flags);
 return NULL;
}

 

static struct page *__rmqueue(struct zone *zone, unsigned int order,
      int migratetype)
{
 struct page *page;

retry_reserve:
 page = __rmqueue_smallest(zone, order, migratetype);

 if (unlikely(!page) && migratetype != MIGRATE_RESERVE) {
  page = __rmqueue_fallback(zone, order, migratetype);

  /*
   * Use MIGRATE_RESERVE rather than fail an allocation. goto
   * is used because __rmqueue_smallest is an inline function
   * and we want just one call site
   */
  if (!page) {
   migratetype = MIGRATE_RESERVE;
   goto retry_reserve;
  }
 }

 trace_mm_page_alloc_zone_locked(page, order, migratetype);
 return page;
}

 

static inline
struct page *__rmqueue_smallest(struct zone *zone, unsigned int order,
      int migratetype)
{
 unsigned int current_order;
 struct free_area * area;
 struct page *page;

 /* Find a page of the appropriate size in the preferred list */
 for (current_order = order; current_order < MAX_ORDER; ++current_order) {
  area = &(zone->free_area[current_order]);
  if (list_empty(&area->free_list[migratetype]))
   continue;

  page = list_entry(area->free_list[migratetype].next,
       struct page, lru);
  list_del(&page->lru);
  rmv_page_order(page);
  area->nr_free--;
  expand(zone, page, order, current_order, area, migratetype);
  return page;
 }

 return NULL;
}

static inline void expand(struct zone *zone, struct page *page,
 int low, int high, struct free_area *area,
 int migratetype)
{
 unsigned long size = 1 << high;

 while (high > low) {
  area--;
  high--;
  size >>= 1;
  VM_BUG_ON(bad_range(zone, &page[size]));
  list_add(&page[size].lru, &area->free_list[migratetype]);
  area->nr_free++;
  set_page_order(&page[size], high);
 }
}

阅读(589) | 评论(0) | 转发(0) |
0

上一篇: kmalloc分配大小的限制

下一篇:vmalloc代码

给主人留下些什么吧!~~