1130: /* Insert into the list of all pools. */
1131: if (__predict_true(!cold))
1132: mutex_enter(&pool_head_lock);
1140: TAILQ_INSERT_BEFORE(pp1, pp, pr_poollist);
1141: if (__predict_true(!cold))
1142: mutex_exit(&pool_head_lock);
1144: /* Insert this into the list of pools using this allocator. */
1145: if (__predict_true(!cold))
1146: mutex_enter(&palloc->pa_lock);
1147: TAILQ_INSERT_TAIL(&palloc->pa_list, pp, pr_alloc_list);
1148: if (__predict_true(!cold))
1149: mutex_exit(&palloc->pa_lock);
1674: cp = pool_allocator_alloc(pp, flags);
1675: if (__predict_true(cp != NULL)) {
1676: ph = pool_alloc_item_header(pp, cp, flags);
2468: /* Add to list of all pools. */
2469: if (__predict_true(!cold))
2470: mutex_enter(&pool_head_lock);
2478: TAILQ_INSERT_BEFORE(pc1, pc, pc_cachelist);
2479: if (__predict_true(!cold))
2480: mutex_exit(&pool_head_lock);
2869:
2870: if (__predict_true((pcg = pc->pc_fullgroups) != NULL)) {
2871: /*
2875: */
2876: if (__predict_true((cur = cc->cc_current) != &pcg_dummy)) {
2877: KASSERT(cur->pcg_avail == 0);
2956: pcg = cc->cc_current;
2957: if (__predict_true(pcg->pcg_avail > 0)) {
2958: object = pcg->pcg_objects[--pcg->pcg_avail].pcgo_va;
2976: pcg = cc->cc_previous;
2977: if (__predict_true(pcg->pcg_avail > 0)) {
2978: cc->cc_previous = cc->cc_current;
3016: if (__predict_false(pc->pc_emptygroups == NULL)) {
3017: if (__predict_true(!pool_cache_disable)) {
3018: pcg = pool_get(pc->pc_pcgpool, PR_NOWAIT);
3019: }
3020: if (__predict_true(pcg != NULL)) {
3021: pcg->pcg_avail = 0;
3066: cur = cc->cc_current;
3067: if (__predict_true(cur != &pcg_dummy)) {
3068: KASSERT(cur->pcg_avail == cur->pcg_size);
3118: pcg = cc->cc_current;
3119: if (__predict_true(pcg->pcg_avail < pcg->pcg_size)) {
3120: pcg->pcg_objects[pcg->pcg_avail].pcgo_va = object;
3132: pcg = cc->cc_previous;
3133: if (__predict_true(pcg->pcg_avail < pcg->pcg_size)) {
3134: cc->cc_previous = cc->cc_current;