415: }
416: if (__predict_false((l->l_flag & LW_IDLE) != 0)) {
417: /* Can't preempt idle loop, don't count as failure. */
420: }
421: if (__predict_false(l->l_nopreempt != 0)) {
422: /* LWP holds preemption disabled, explicitly. */
428: }
429: if (__predict_false((l->l_pflag & LP_INTR) != 0)) {
430: /* Can't preempt soft interrupts yet. */
435: s = splsched();
436: if (__predict_false(l->l_blcnt != 0 ||
437: curcpu()->ci_biglock_wanted != NULL)) {
445: }
446: if (__predict_false(!cpu_kpreempt_enter(where, s))) {
447: /*
470: /* Record preemption failure for reporting via lockstat. */
471: if (__predict_false(failed)) {
472: int lsflag = 0;
475: /* Might recurse, make it atomic. */
476: if (__predict_false(lsflag)) {
477: if (where == 0) {
704: l->l_dopreempt = 0;
705: if (__predict_false(l->l_pfailaddr != 0)) {
706: LOCKSTAT_FLAG(lsflag);
1181: PROCLIST_FOREACH(p, &allproc) {
1182: if (__predict_false((p->p_flag & PK_MARKER) != 0))
1183: continue;
1191: LIST_FOREACH(l, &p->p_lwps, l_sibling) {
1192: if (__predict_false((l->l_flag & LW_IDLE) != 0))
1193: continue;
1218: sig = 0;
1219: if (__predict_false(runtm >= rlim->rlim_cur)) {
1220: if (runtm >= rlim->rlim_max)
1228: mutex_exit(p->p_lock);
1229: if (__predict_false(runtm < 0)) {
1230: if (!backwards) {
1234: }
1235: } else if (__predict_false(sig)) {
1236: KASSERT((p->p_flag & PK_SYSTEM) == 0);