lp 127 crypto/ccm.c unsigned int lp = req->iv[0];
lp 128 crypto/ccm.c unsigned int l = lp + 1;
lp 846 fs/gfs2/dir.c __be64 *lp;
lp 903 fs/gfs2/dir.c lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode));
lp 905 fs/gfs2/dir.c for (x = sdp->sd_hash_ptrs; x--; lp++)
lp 906 fs/gfs2/dir.c *lp = cpu_to_be64(bn);
lp 939 fs/gfs2/dir.c __be64 *lp;
lp 984 fs/gfs2/dir.c lp = kmalloc(half_len * sizeof(__be64), GFP_NOFS | __GFP_NOFAIL);
lp 987 fs/gfs2/dir.c lp[x] = cpu_to_be64(bn);
lp 989 fs/gfs2/dir.c error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64),
lp 997 fs/gfs2/dir.c kfree(lp);
lp 1058 fs/gfs2/dir.c kfree(lp);
lp 1367 fs/gfs2/dir.c __be64 *lp;
lp 1381 fs/gfs2/dir.c lp = kmalloc(sdp->sd_hash_bsize, GFP_NOFS);
lp 1382 fs/gfs2/dir.c if (!lp)
lp 1390 fs/gfs2/dir.c error = gfs2_dir_read_data(dip, (char *)lp,
lp 1403 fs/gfs2/dir.c be64_to_cpu(lp[lp_offset]));
lp 1412 fs/gfs2/dir.c kfree(lp);
lp 1782 fs/gfs2/dir.c __be64 *lp;
lp 1792 fs/gfs2/dir.c lp = kmalloc(sdp->sd_hash_bsize, GFP_NOFS);
lp 1793 fs/gfs2/dir.c if (!lp)
lp 1801 fs/gfs2/dir.c error = gfs2_dir_read_data(dip, (char *)lp,
lp 1812 fs/gfs2/dir.c leaf_no = be64_to_cpu(lp[lp_offset]);
lp 1836 fs/gfs2/dir.c kfree(lp);
lp 39 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 42 fs/gfs2/locking/dlm/lock.c list_add_tail(&lp->delay_list, &ls->submit);
lp 49 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_AST_WAIT, &lp->flags);
lp 51 fs/gfs2/locking/dlm/lock.c wake_up_bit(&lp->flags, LFL_AST_WAIT);
lp 56 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 59 fs/gfs2/locking/dlm/lock.c if (!list_empty(&lp->delay_list))
lp 60 fs/gfs2/locking/dlm/lock.c list_del_init(&lp->delay_list);
lp 64 fs/gfs2/locking/dlm/lock.c kfree(lp);
lp 69 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 72 fs/gfs2/locking/dlm/lock.c list_add_tail(&lp->delay_list, &ls->delayed);
lp 78 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 83 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_status == -DLM_ECANCEL) {
lp 85 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 86 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number,
lp 87 fs/gfs2/locking/dlm/lock.c lp->flags);
lp 89 fs/gfs2/locking/dlm/lock.c lp->req = lp->cur;
lp 91 fs/gfs2/locking/dlm/lock.c if (lp->cur == DLM_LOCK_IV)
lp 92 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lkid = 0;
lp 96 fs/gfs2/locking/dlm/lock.c if (test_and_clear_bit(LFL_DLM_UNLOCK, &lp->flags)) {
lp 97 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_status != -DLM_EUNLOCK) {
lp 99 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_status, lp->lockname.ln_type,
lp 100 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number,
lp 101 fs/gfs2/locking/dlm/lock.c lp->flags);
lp 105 fs/gfs2/locking/dlm/lock.c lp->cur = DLM_LOCK_IV;
lp 106 fs/gfs2/locking/dlm/lock.c lp->req = DLM_LOCK_IV;
lp 107 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lkid = 0;
lp 109 fs/gfs2/locking/dlm/lock.c if (test_and_clear_bit(LFL_UNLOCK_DELETE, &lp->flags)) {
lp 110 fs/gfs2/locking/dlm/lock.c gdlm_delete_lp(lp);
lp 116 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_flags & DLM_SBF_VALNOTVALID)
lp 117 fs/gfs2/locking/dlm/lock.c memset(lp->lksb.sb_lvbptr, 0, GDLM_LVB_SIZE);
lp 119 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_flags & DLM_SBF_ALTMODE) {
lp 120 fs/gfs2/locking/dlm/lock.c if (lp->req == DLM_LOCK_PR)
lp 121 fs/gfs2/locking/dlm/lock.c lp->req = DLM_LOCK_CW;
lp 122 fs/gfs2/locking/dlm/lock.c else if (lp->req == DLM_LOCK_CW)
lp 123 fs/gfs2/locking/dlm/lock.c lp->req = DLM_LOCK_PR;
lp 131 fs/gfs2/locking/dlm/lock.c if (test_and_clear_bit(LFL_CANCEL, &lp->flags)) {
lp 133 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 134 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number);
lp 135 fs/gfs2/locking/dlm/lock.c lp->req = lp->cur;
lp 144 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_status) {
lp 146 fs/gfs2/locking/dlm/lock.c if ((lp->lksb.sb_status == -EAGAIN) &&
lp 147 fs/gfs2/locking/dlm/lock.c (lp->lkf & DLM_LKF_NOQUEUE)) {
lp 148 fs/gfs2/locking/dlm/lock.c lp->req = lp->cur;
lp 149 fs/gfs2/locking/dlm/lock.c if (lp->cur == DLM_LOCK_IV)
lp 150 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lkid = 0;
lp 156 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_status, lp->lockname.ln_type,
lp 157 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number,
lp 158 fs/gfs2/locking/dlm/lock.c lp->flags);
lp 166 fs/gfs2/locking/dlm/lock.c if (test_and_clear_bit(LFL_SYNC_LVB, &lp->flags)) {
lp 167 fs/gfs2/locking/dlm/lock.c wake_up_ast(lp);
lp 177 fs/gfs2/locking/dlm/lock.c if (test_and_clear_bit(LFL_REREQUEST, &lp->flags)) {
lp 178 fs/gfs2/locking/dlm/lock.c gdlm_assert(lp->req == DLM_LOCK_NL, "%x,%llx",
lp 179 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 180 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number);
lp 181 fs/gfs2/locking/dlm/lock.c gdlm_assert(lp->prev_req > DLM_LOCK_NL, "%x,%llx",
lp 182 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 183 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number);
lp 185 fs/gfs2/locking/dlm/lock.c lp->cur = DLM_LOCK_NL;
lp 186 fs/gfs2/locking/dlm/lock.c lp->req = lp->prev_req;
lp 187 fs/gfs2/locking/dlm/lock.c lp->prev_req = DLM_LOCK_IV;
lp 188 fs/gfs2/locking/dlm/lock.c lp->lkf &= ~DLM_LKF_CONVDEADLK;
lp 190 fs/gfs2/locking/dlm/lock.c set_bit(LFL_NOCACHE, &lp->flags);
lp 193 fs/gfs2/locking/dlm/lock.c !test_bit(LFL_NOBLOCK, &lp->flags))
lp 194 fs/gfs2/locking/dlm/lock.c gdlm_queue_delayed(lp);
lp 196 fs/gfs2/locking/dlm/lock.c queue_submit(lp);
lp 210 fs/gfs2/locking/dlm/lock.c !test_bit(LFL_NOBLOCK, &lp->flags) &&
lp 211 fs/gfs2/locking/dlm/lock.c lp->req != DLM_LOCK_NL) {
lp 213 fs/gfs2/locking/dlm/lock.c lp->cur = lp->req;
lp 214 fs/gfs2/locking/dlm/lock.c lp->prev_req = lp->req;
lp 215 fs/gfs2/locking/dlm/lock.c lp->req = DLM_LOCK_NL;
lp 216 fs/gfs2/locking/dlm/lock.c lp->lkf |= DLM_LKF_CONVERT;
lp 217 fs/gfs2/locking/dlm/lock.c lp->lkf &= ~DLM_LKF_CONVDEADLK;
lp 220 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 221 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number,
lp 222 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lkid, lp->cur, lp->req);
lp 224 fs/gfs2/locking/dlm/lock.c set_bit(LFL_REREQUEST, &lp->flags);
lp 225 fs/gfs2/locking/dlm/lock.c queue_submit(lp);
lp 234 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_flags & DLM_SBF_DEMOTED)
lp 235 fs/gfs2/locking/dlm/lock.c set_bit(LFL_NOCACHE, &lp->flags);
lp 242 fs/gfs2/locking/dlm/lock.c if (test_bit(LFL_INLOCK, &lp->flags)) {
lp 243 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_NOBLOCK, &lp->flags);
lp 244 fs/gfs2/locking/dlm/lock.c lp->cur = lp->req;
lp 245 fs/gfs2/locking/dlm/lock.c wake_up_ast(lp);
lp 253 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_NOBLOCK, &lp->flags);
lp 254 fs/gfs2/locking/dlm/lock.c lp->cur = lp->req;
lp 256 fs/gfs2/locking/dlm/lock.c acb.lc_name = lp->lockname;
lp 257 fs/gfs2/locking/dlm/lock.c acb.lc_ret |= gdlm_make_lmstate(lp->cur);
lp 264 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = astarg;
lp 265 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_ACTIVE, &lp->flags);
lp 266 fs/gfs2/locking/dlm/lock.c process_complete(lp);
lp 271 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 288 fs/gfs2/locking/dlm/lock.c ls->fscb(ls->sdp, cb, &lp->lockname);
lp 294 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = astarg;
lp 298 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 299 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number);
lp 303 fs/gfs2/locking/dlm/lock.c process_blocking(lp, mode);
lp 331 fs/gfs2/locking/dlm/lock.c if (lp->cur != DLM_LOCK_IV)
lp 332 fs/gfs2/locking/dlm/lock.c gdlm_assert(lp->cur == cur, "%d, %d", lp->cur, cur);
lp 361 fs/gfs2/locking/dlm/lock.c if (lp->lksb.sb_lkid != 0) {
lp 365 fs/gfs2/locking/dlm/lock.c if (lp->lvb)
lp 384 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp;
lp 386 fs/gfs2/locking/dlm/lock.c lp = kzalloc(sizeof(struct gdlm_lock), GFP_NOFS);
lp 387 fs/gfs2/locking/dlm/lock.c if (!lp)
lp 390 fs/gfs2/locking/dlm/lock.c lp->lockname = *name;
lp 391 fs/gfs2/locking/dlm/lock.c make_strname(name, &lp->strname);
lp 392 fs/gfs2/locking/dlm/lock.c lp->ls = ls;
lp 393 fs/gfs2/locking/dlm/lock.c lp->cur = DLM_LOCK_IV;
lp 394 fs/gfs2/locking/dlm/lock.c INIT_LIST_HEAD(&lp->delay_list);
lp 400 fs/gfs2/locking/dlm/lock.c *lpp = lp;
lp 407 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp;
lp 410 fs/gfs2/locking/dlm/lock.c error = gdlm_create_lp(lockspace, name, &lp);
lp 412 fs/gfs2/locking/dlm/lock.c *lockp = lp;
lp 423 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 433 fs/gfs2/locking/dlm/lock.c !test_bit(LFL_NOBLOCK, &lp->flags) && lp->req != DLM_LOCK_NL) {
lp 434 fs/gfs2/locking/dlm/lock.c gdlm_queue_delayed(lp);
lp 442 fs/gfs2/locking/dlm/lock.c if (test_bit(LFL_NOBAST, &lp->flags))
lp 445 fs/gfs2/locking/dlm/lock.c set_bit(LFL_ACTIVE, &lp->flags);
lp 447 fs/gfs2/locking/dlm/lock.c log_debug("lk %x,%llx id %x %d,%d %x", lp->lockname.ln_type,
lp 448 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, lp->lksb.sb_lkid,
lp 449 fs/gfs2/locking/dlm/lock.c lp->cur, lp->req, lp->lkf);
lp 451 fs/gfs2/locking/dlm/lock.c error = dlm_lock(ls->dlm_lockspace, lp->req, &lp->lksb, lp->lkf,
lp 452 fs/gfs2/locking/dlm/lock.c lp->strname.name, lp->strname.namelen, 0, gdlm_ast,
lp 453 fs/gfs2/locking/dlm/lock.c lp, bast ? gdlm_bast : NULL);
lp 455 fs/gfs2/locking/dlm/lock.c if ((error == -EAGAIN) && (lp->lkf & DLM_LKF_NOQUEUE)) {
lp 456 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_status = -EAGAIN;
lp 457 fs/gfs2/locking/dlm/lock.c gdlm_ast(lp);
lp 463 fs/gfs2/locking/dlm/lock.c "flags=%lx", ls->fsname, lp->lockname.ln_type,
lp 464 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, error,
lp 465 fs/gfs2/locking/dlm/lock.c lp->cur, lp->req, lp->lkf, lp->flags);
lp 473 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 477 fs/gfs2/locking/dlm/lock.c set_bit(LFL_DLM_UNLOCK, &lp->flags);
lp 478 fs/gfs2/locking/dlm/lock.c set_bit(LFL_ACTIVE, &lp->flags);
lp 480 fs/gfs2/locking/dlm/lock.c if (lp->lvb)
lp 483 fs/gfs2/locking/dlm/lock.c log_debug("un %x,%llx %x %d %x", lp->lockname.ln_type,
lp 484 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number,
lp 485 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lkid, lp->cur, lkf);
lp 487 fs/gfs2/locking/dlm/lock.c error = dlm_unlock(ls->dlm_lockspace, lp->lksb.sb_lkid, lkf, NULL, lp);
lp 491 fs/gfs2/locking/dlm/lock.c "flags=%lx", ls->fsname, lp->lockname.ln_type,
lp 492 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, error,
lp 493 fs/gfs2/locking/dlm/lock.c lp->cur, lp->req, lp->lkf, lp->flags);
lp 502 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = lock;
lp 510 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_DLM_CANCEL, &lp->flags);
lp 512 fs/gfs2/locking/dlm/lock.c set_bit(LFL_NOBLOCK, &lp->flags);
lp 514 fs/gfs2/locking/dlm/lock.c check_cur_state(lp, cur_state);
lp 515 fs/gfs2/locking/dlm/lock.c lp->req = make_mode(req_state);
lp 516 fs/gfs2/locking/dlm/lock.c lp->lkf = make_flags(lp, flags, lp->cur, lp->req);
lp 518 fs/gfs2/locking/dlm/lock.c return gdlm_do_lock(lp);
lp 523 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = lock;
lp 525 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_DLM_CANCEL, &lp->flags);
lp 526 fs/gfs2/locking/dlm/lock.c if (lp->cur == DLM_LOCK_IV)
lp 528 fs/gfs2/locking/dlm/lock.c return gdlm_do_unlock(lp);
lp 533 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = lock;
lp 534 fs/gfs2/locking/dlm/lock.c struct gdlm_ls *ls = lp->ls;
lp 537 fs/gfs2/locking/dlm/lock.c if (test_bit(LFL_DLM_CANCEL, &lp->flags))
lp 540 fs/gfs2/locking/dlm/lock.c log_info("gdlm_cancel %x,%llx flags %lx", lp->lockname.ln_type,
lp 541 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, lp->flags);
lp 544 fs/gfs2/locking/dlm/lock.c if (!list_empty(&lp->delay_list)) {
lp 545 fs/gfs2/locking/dlm/lock.c list_del_init(&lp->delay_list);
lp 551 fs/gfs2/locking/dlm/lock.c set_bit(LFL_CANCEL, &lp->flags);
lp 552 fs/gfs2/locking/dlm/lock.c set_bit(LFL_ACTIVE, &lp->flags);
lp 553 fs/gfs2/locking/dlm/lock.c gdlm_ast(lp);
lp 557 fs/gfs2/locking/dlm/lock.c if (!test_bit(LFL_ACTIVE, &lp->flags) ||
lp 558 fs/gfs2/locking/dlm/lock.c test_bit(LFL_DLM_UNLOCK, &lp->flags)) {
lp 560 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 561 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, lp->flags);
lp 567 fs/gfs2/locking/dlm/lock.c set_bit(LFL_DLM_CANCEL, &lp->flags);
lp 568 fs/gfs2/locking/dlm/lock.c set_bit(LFL_ACTIVE, &lp->flags);
lp 570 fs/gfs2/locking/dlm/lock.c error = dlm_unlock(ls->dlm_lockspace, lp->lksb.sb_lkid, DLM_LKF_CANCEL,
lp 571 fs/gfs2/locking/dlm/lock.c NULL, lp);
lp 574 fs/gfs2/locking/dlm/lock.c lp->lockname.ln_type,
lp 575 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number, lp->flags);
lp 578 fs/gfs2/locking/dlm/lock.c clear_bit(LFL_DLM_CANCEL, &lp->flags);
lp 589 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lvbptr = lvb;
lp 590 fs/gfs2/locking/dlm/lock.c lp->lvb = lvb;
lp 596 fs/gfs2/locking/dlm/lock.c kfree(lp->lvb);
lp 597 fs/gfs2/locking/dlm/lock.c lp->lvb = NULL;
lp 598 fs/gfs2/locking/dlm/lock.c lp->lksb.sb_lvbptr = NULL;
lp 616 fs/gfs2/locking/dlm/lock.c if (lp->hold_null) {
lp 621 fs/gfs2/locking/dlm/lock.c error = gdlm_create_lp(lp->ls, &lp->lockname, &lpn);
lp 644 fs/gfs2/locking/dlm/lock.c lp->hold_null = lpn;
lp 655 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lpn = lp->hold_null;
lp 657 fs/gfs2/locking/dlm/lock.c gdlm_assert(lpn, "%x,%llx", lp->lockname.ln_type,
lp 658 fs/gfs2/locking/dlm/lock.c (unsigned long long)lp->lockname.ln_number);
lp 663 fs/gfs2/locking/dlm/lock.c lp->hold_null = NULL;
lp 672 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = lock;
lp 675 fs/gfs2/locking/dlm/lock.c error = gdlm_add_lvb(lp);
lp 679 fs/gfs2/locking/dlm/lock.c *lvbp = lp->lvb;
lp 681 fs/gfs2/locking/dlm/lock.c error = hold_null_lock(lp);
lp 683 fs/gfs2/locking/dlm/lock.c gdlm_del_lvb(lp);
lp 690 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp = lock;
lp 692 fs/gfs2/locking/dlm/lock.c unhold_null_lock(lp);
lp 693 fs/gfs2/locking/dlm/lock.c gdlm_del_lvb(lp);
lp 698 fs/gfs2/locking/dlm/lock.c struct gdlm_lock *lp, *safe;
lp 701 fs/gfs2/locking/dlm/lock.c list_for_each_entry_safe(lp, safe, &ls->delayed, delay_list) {
lp 702 fs/gfs2/locking/dlm/lock.c list_del_init(&lp->delay_list);
lp 703 fs/gfs2/locking/dlm/lock.c list_add_tail(&lp->delay_list, &ls->submit);
lp 26 fs/gfs2/locking/dlm/thread.c struct gdlm_lock *lp = NULL;
lp 35 fs/gfs2/locking/dlm/thread.c lp = list_entry(ls->submit.next, struct gdlm_lock,
lp 37 fs/gfs2/locking/dlm/thread.c list_del_init(&lp->delay_list);
lp 39 fs/gfs2/locking/dlm/thread.c gdlm_do_lock(lp);
lp 2795 fs/jfs/jfs_dmap.c int lp, pp, k;
lp 2800 fs/jfs/jfs_dmap.c lp = leafno + le32_to_cpu(tp->dmt_leafidx);
lp 2805 fs/jfs/jfs_dmap.c if (tp->dmt_stree[lp] == newval)
lp 2810 fs/jfs/jfs_dmap.c tp->dmt_stree[lp] = newval;
lp 2818 fs/jfs/jfs_dmap.c lp = ((lp - 1) & ~0x03) + 1;
lp 2822 fs/jfs/jfs_dmap.c pp = (lp - 1) >> 2;
lp 2826 fs/jfs/jfs_dmap.c max = TREEMAX(&tp->dmt_stree[lp]);
lp 2840 fs/jfs/jfs_dmap.c lp = pp;
lp 940 fs/jfs/jfs_dtree.c dtpage_t *lp; /* left child page */
lp 1132 fs/jfs/jfs_dtree.c lp = sp;
lp 1186 fs/jfs/jfs_dtree.c rc = ciGetLeafPrefixKey(lp,
lp 1187 fs/jfs/jfs_dtree.c lp->header.nextindex-1,
lp 2447 fs/jfs/jfs_dtree.c dtpage_t *p, *pp, *rp = 0, *lp= 0;
lp 2507 fs/jfs/jfs_dtree.c DT_GETPAGE(ip, prevbn, lmp, PSIZE, lp, rc);
lp 2532 fs/jfs/jfs_dtree.c lp->header.next = cpu_to_le64(nxaddr);
lp 3785 fs/jfs/jfs_dtree.c dtGetKey(lp, li, &lkey, flag);
lp 360 fs/jfs/jfs_logmgr.c struct logpage *lp; /* dst log page */
lp 379 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 430 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 443 fs/jfs/jfs_logmgr.c dst = (caddr_t) lp + dstoffset;
lp 455 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 466 fs/jfs/jfs_logmgr.c lvd = (struct lvd *) ((caddr_t) lp + dstoffset);
lp 491 fs/jfs/jfs_logmgr.c dst = (caddr_t) lp + dstoffset;
lp 551 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 573 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 583 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 584 fs/jfs/jfs_logmgr.c lspn = le32_to_cpu(lp->h.page);
lp 634 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
lp 654 fs/jfs/jfs_logmgr.c lp = (struct logpage *) nextbp->l_ldata;
lp 655 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn + 1);
lp 656 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE);
lp 759 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 788 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 795 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
lp 803 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor);
lp 825 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 885 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 887 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
lp 949 fs/jfs/jfs_logmgr.c struct logsyncblk *lp;
lp 970 fs/jfs/jfs_logmgr.c lp = list_entry(log->synclist.next,
lp 972 fs/jfs/jfs_logmgr.c log->sync = lp->lsn;
lp 1274 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 1366 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 1370 fs/jfs/jfs_logmgr.c le16_to_cpu(lp->h.eor));
lp 1398 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 1399 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
lp 1622 fs/jfs/jfs_logmgr.c struct logsyncblk *lp;
lp 1625 fs/jfs/jfs_logmgr.c list_for_each_entry(lp, &log->synclist, synclist) {
lp 1626 fs/jfs/jfs_logmgr.c if (lp->xflag & COMMIT_PAGE) {
lp 1627 fs/jfs/jfs_logmgr.c struct metapage *mp = (struct metapage *)lp;
lp 1638 fs/jfs/jfs_logmgr.c lp, sizeof(struct tblock), 0);
lp 1669 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 1687 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 1688 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor);
lp 2393 fs/jfs/jfs_logmgr.c struct logpage *lp;
lp 2460 fs/jfs/jfs_logmgr.c lp = (struct logpage *) bp->l_ldata;
lp 2465 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(npages - 3);
lp 2466 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE + LOGRDSIZE);
lp 2468 fs/jfs/jfs_logmgr.c lrd_ptr = (struct lrd *) &lp->data;
lp 2485 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn);
lp 2486 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE);
lp 2829 fs/jfs/jfs_xtree.c xtpage_t *p, *pp, *rp, *lp; /* base B+-tree index page */
lp 3003 fs/jfs/jfs_xtree.c XT_GETPAGE(ip, prevbn, lmp, PSIZE, lp, rc);
lp 3021 fs/jfs/jfs_xtree.c lp->header.next = cpu_to_le64(nxaddr);
lp 627 fs/ubifs/debug.c "flags %#x\n", lp->lnum, lp->free, lp->dirty,
lp 628 fs/ubifs/debug.c c->leb_size - lp->free - lp->dirty, lp->flags);
lp 634 fs/ubifs/debug.c struct ubifs_lprops lp;
lp 642 fs/ubifs/debug.c err = ubifs_read_one_lp(c, lnum, &lp);
lp 646 fs/ubifs/debug.c dbg_dump_lprop(c, &lp);
lp 745 fs/ubifs/debug.c struct ubifs_lprops *lp = &pnode->lprops[i];
lp 748 fs/ubifs/debug.c i, lp->free, lp->dirty, lp->flags, lp->lnum);
lp 237 fs/ubifs/find.c const struct ubifs_lprops *lp = NULL, *idx_lp = NULL;
lp 264 fs/ubifs/find.c lp = ubifs_fast_find_empty(c);
lp 265 fs/ubifs/find.c if (lp)
lp 269 fs/ubifs/find.c lp = ubifs_fast_find_freeable(c);
lp 270 fs/ubifs/find.c if (lp)
lp 305 fs/ubifs/find.c lp = heap->arr[0];
lp 306 fs/ubifs/find.c if (lp->dirty + lp->free < min_space)
lp 307 fs/ubifs/find.c lp = NULL;
lp 311 fs/ubifs/find.c if (idx_lp && lp) {
lp 312 fs/ubifs/find.c if (idx_lp->free + idx_lp->dirty >= lp->free + lp->dirty)
lp 313 fs/ubifs/find.c lp = idx_lp;
lp 314 fs/ubifs/find.c } else if (idx_lp && !lp)
lp 315 fs/ubifs/find.c lp = idx_lp;
lp 317 fs/ubifs/find.c if (lp) {
lp 318 fs/ubifs/find.c ubifs_assert(lp->free + lp->dirty >= c->dead_wm);
lp 324 fs/ubifs/find.c lp = scan_for_dirty(c, min_space, pick_free, exclude_index);
lp 325 fs/ubifs/find.c if (IS_ERR(lp)) {
lp 326 fs/ubifs/find.c err = PTR_ERR(lp);
lp 329 fs/ubifs/find.c ubifs_assert(lp->dirty >= c->dead_wm ||
lp 330 fs/ubifs/find.c (pick_free && lp->free + lp->dirty == c->leb_size));
lp 334 fs/ubifs/find.c lp->lnum, lp->free, lp->dirty, lp->flags);
lp 336 fs/ubifs/find.c lp = ubifs_change_lp(c, lp, LPROPS_NC, LPROPS_NC,
lp 337 fs/ubifs/find.c lp->flags | LPROPS_TAKEN, 0);
lp 338 fs/ubifs/find.c if (IS_ERR(lp)) {
lp 339 fs/ubifs/find.c err = PTR_ERR(lp);
lp 343 fs/ubifs/find.c memcpy(ret_lp, lp, sizeof(struct ubifs_lprops));
lp 892 fs/ubifs/find.c const struct ubifs_lprops *lp;
lp 903 fs/ubifs/find.c lp = ubifs_lpt_lookup_dirty(c, lnum);
lp 904 fs/ubifs/find.c if (unlikely(IS_ERR(lp)))
lp 905 fs/ubifs/find.c return PTR_ERR(lp);
lp 906 fs/ubifs/find.c lp = ubifs_change_lp(c, lp, LPROPS_NC, LPROPS_NC,
lp 907 fs/ubifs/find.c lp->flags | LPROPS_INDEX, -1);
lp 908 fs/ubifs/find.c if (unlikely(IS_ERR(lp)))
lp 909 fs/ubifs/find.c return PTR_ERR(lp);
lp 911 fs/ubifs/find.c lp->lnum, lp->dirty, lp->free, lp->flags);
lp 921 fs/ubifs/find.c const struct ubifs_lprops *lp;
lp 929 fs/ubifs/find.c lp = ubifs_lpt_lookup(c, lnum);
lp 930 fs/ubifs/find.c if (IS_ERR(lp))
lp 931 fs/ubifs/find.c return PTR_ERR(lp);
lp 932 fs/ubifs/find.c if ((lp->flags & LPROPS_TAKEN) || !(lp->flags & LPROPS_INDEX))
lp 934 fs/ubifs/find.c lp = ubifs_change_lp(c, lp, LPROPS_NC, LPROPS_NC,
lp 935 fs/ubifs/find.c lp->flags | LPROPS_TAKEN, 0);
lp 936 fs/ubifs/find.c if (IS_ERR(lp))
lp 937 fs/ubifs/find.c return PTR_ERR(lp);
lp 940 fs/ubifs/find.c dbg_find("LEB %d, dirty %d and free %d flags %#x", lp->lnum, lp->dirty,
lp 941 fs/ubifs/find.c lp->free, lp->flags);
lp 942 fs/ubifs/find.c ubifs_assert(lp->flags | LPROPS_TAKEN);
lp 943 fs/ubifs/find.c ubifs_assert(lp->flags | LPROPS_INDEX);
lp 275 fs/ubifs/gc.c int err = 0, lnum = lp->lnum;
lp 297 fs/ubifs/gc.c lnum, lp->free, lp->dirty);
lp 333 fs/ubifs/gc.c lnum, lp->free, lp->dirty);
lp 421 fs/ubifs/gc.c struct ubifs_lprops lp;
lp 479 fs/ubifs/gc.c ret = ubifs_find_dirty_leb(c, &lp, min_space, anyway ? 0 : 1);
lp 487 fs/ubifs/gc.c "(min. space %d)", lp.lnum, lp.free, lp.dirty,
lp 488 fs/ubifs/gc.c lp.free + lp.dirty, min_space);
lp 490 fs/ubifs/gc.c if (lp.free + lp.dirty == c->leb_size) {
lp 492 fs/ubifs/gc.c dbg_gc("LEB %d is free, return it", lp.lnum);
lp 497 fs/ubifs/gc.c ubifs_assert(!(lp.flags & LPROPS_INDEX));
lp 498 fs/ubifs/gc.c if (lp.free != c->leb_size) {
lp 508 fs/ubifs/gc.c ret = ubifs_change_one_lp(c, lp.lnum,
lp 514 fs/ubifs/gc.c ret = ubifs_leb_unmap(c, lp.lnum);
lp 517 fs/ubifs/gc.c ret = lp.lnum;
lp 525 fs/ubifs/gc.c ret = ubifs_garbage_collect_leb(c, &lp);
lp 536 fs/ubifs/gc.c err = ubifs_return_leb(c, lp.lnum);
lp 546 fs/ubifs/gc.c dbg_gc("LEB %d freed, return", lp.lnum);
lp 547 fs/ubifs/gc.c ret = lp.lnum;
lp 558 fs/ubifs/gc.c dbg_gc("indexing LEB %d freed, continue", lp.lnum);
lp 564 fs/ubifs/gc.c dbg_gc("LEB %d retained, freed %d bytes", lp.lnum,
lp 627 fs/ubifs/gc.c ubifs_return_leb(c, lp.lnum);
lp 645 fs/ubifs/gc.c const struct ubifs_lprops *lp;
lp 655 fs/ubifs/gc.c lp = ubifs_fast_find_freeable(c);
lp 656 fs/ubifs/gc.c if (unlikely(IS_ERR(lp))) {
lp 657 fs/ubifs/gc.c err = PTR_ERR(lp);
lp 660 fs/ubifs/gc.c if (!lp)
lp 662 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_TAKEN));
lp 663 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_INDEX));
lp 664 fs/ubifs/gc.c err = ubifs_leb_unmap(c, lp->lnum);
lp 667 fs/ubifs/gc.c lp = ubifs_change_lp(c, lp, c->leb_size, 0, lp->flags, 0);
lp 668 fs/ubifs/gc.c if (unlikely(IS_ERR(lp))) {
lp 669 fs/ubifs/gc.c err = PTR_ERR(lp);
lp 672 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_TAKEN));
lp 673 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_INDEX));
lp 682 fs/ubifs/gc.c lp = ubifs_fast_find_frdi_idx(c);
lp 683 fs/ubifs/gc.c if (unlikely(IS_ERR(lp))) {
lp 684 fs/ubifs/gc.c err = PTR_ERR(lp);
lp 687 fs/ubifs/gc.c if (!lp)
lp 694 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_TAKEN));
lp 695 fs/ubifs/gc.c ubifs_assert(lp->flags & LPROPS_INDEX);
lp 697 fs/ubifs/gc.c flags = (lp->flags | LPROPS_TAKEN) ^ LPROPS_INDEX;
lp 698 fs/ubifs/gc.c lp = ubifs_change_lp(c, lp, c->leb_size, 0, flags, 1);
lp 699 fs/ubifs/gc.c if (unlikely(IS_ERR(lp))) {
lp 700 fs/ubifs/gc.c err = PTR_ERR(lp);
lp 704 fs/ubifs/gc.c ubifs_assert(lp->flags & LPROPS_TAKEN);
lp 705 fs/ubifs/gc.c ubifs_assert(!(lp->flags & LPROPS_INDEX));
lp 706 fs/ubifs/gc.c idx_gc->lnum = lp->lnum;
lp 196 fs/ubifs/lprops.c struct ubifs_lprops *lp;
lp 198 fs/ubifs/lprops.c lp = heap->arr[cpos];
lp 199 fs/ubifs/lprops.c lp->flags &= ~LPROPS_CAT_MASK;
lp 200 fs/ubifs/lprops.c lp->flags |= LPROPS_UNCAT;
lp 201 fs/ubifs/lprops.c list_add(&lp->list, &c->uncat_list);
lp 551 fs/ubifs/lprops.c struct ubifs_lprops *lprops = (struct ubifs_lprops *)lp;
lp 695 fs/ubifs/lprops.c const struct ubifs_lprops *lp;
lp 699 fs/ubifs/lprops.c lp = ubifs_lpt_lookup_dirty(c, lnum);
lp 700 fs/ubifs/lprops.c if (IS_ERR(lp)) {
lp 701 fs/ubifs/lprops.c err = PTR_ERR(lp);
lp 705 fs/ubifs/lprops.c flags = (lp->flags | flags_set) & ~flags_clean;
lp 706 fs/ubifs/lprops.c lp = ubifs_change_lp(c, lp, free, dirty, flags, idx_gc_cnt);
lp 707 fs/ubifs/lprops.c if (IS_ERR(lp))
lp 708 fs/ubifs/lprops.c err = PTR_ERR(lp);
lp 731 fs/ubifs/lprops.c const struct ubifs_lprops *lp;
lp 735 fs/ubifs/lprops.c lp = ubifs_lpt_lookup_dirty(c, lnum);
lp 736 fs/ubifs/lprops.c if (IS_ERR(lp)) {
lp 737 fs/ubifs/lprops.c err = PTR_ERR(lp);
lp 741 fs/ubifs/lprops.c flags = (lp->flags | flags_set) & ~flags_clean;
lp 742 fs/ubifs/lprops.c lp = ubifs_change_lp(c, lp, free, lp->dirty + dirty, flags, 0);
lp 743 fs/ubifs/lprops.c if (IS_ERR(lp))
lp 744 fs/ubifs/lprops.c err = PTR_ERR(lp);
lp 774 fs/ubifs/lprops.c memcpy(lp, lpp, sizeof(struct ubifs_lprops));
lp 992 fs/ubifs/lprops.c struct ubifs_lprops *lp;
lp 1003 fs/ubifs/lprops.c lp = ubifs_lpt_lookup(c, lprops->lnum);
lp 1004 fs/ubifs/lprops.c if (IS_ERR(lp)) {
lp 1008 fs/ubifs/lprops.c if (lprops != lp) {
lp 1010 fs/ubifs/lprops.c (size_t)lprops, (size_t)lp, lprops->lnum,
lp 1011 fs/ubifs/lprops.c lp->lnum);
lp 1016 fs/ubifs/lprops.c lp = heap->arr[j];
lp 1017 fs/ubifs/lprops.c if (lp == lprops) {
lp 1021 fs/ubifs/lprops.c if (lp->lnum == lprops->lnum) {
lp 1064 fs/ubifs/lprops.c int cat, lnum = lp->lnum, is_idx = 0, used = 0, free, dirty;
lp 1066 fs/ubifs/lprops.c cat = lp->flags & LPROPS_CAT_MASK;
lp 1068 fs/ubifs/lprops.c cat = ubifs_categorize_lprops(c, lp);
lp 1069 fs/ubifs/lprops.c if (cat != (lp->flags & LPROPS_CAT_MASK)) {
lp 1071 fs/ubifs/lprops.c (lp->flags & LPROPS_CAT_MASK), cat);
lp 1099 fs/ubifs/lprops.c if (lprops == lp) {
lp 1115 fs/ubifs/lprops.c if ((lp->hpos != -1 && heap->arr[lp->hpos]->lnum != lnum) ||
lp 1116 fs/ubifs/lprops.c lp != heap->arr[lp->hpos]) {
lp 1128 fs/ubifs/lprops.c if (lp->free == c->leb_size) {
lp 1137 fs/ubifs/lprops.c if (lp->free + lp->dirty == c->leb_size &&
lp 1138 fs/ubifs/lprops.c !(lp->flags & LPROPS_INDEX)) {
lp 1141 fs/ubifs/lprops.c lst->total_free += lp->free;
lp 1142 fs/ubifs/lprops.c lst->total_dirty += lp->dirty;
lp 1191 fs/ubifs/lprops.c if (lp->free + lp->dirty == c->leb_size &&
lp 1193 fs/ubifs/lprops.c if ((is_idx && !(lp->flags & LPROPS_INDEX)) ||
lp 1195 fs/ubifs/lprops.c lp->free == c->leb_size) {
lp 1203 fs/ubifs/lprops.c free = lp->free;
lp 1204 fs/ubifs/lprops.c dirty = lp->dirty;
lp 1208 fs/ubifs/lprops.c if (is_idx && lp->free + lp->dirty == free + dirty &&
lp 1221 fs/ubifs/lprops.c free = lp->free;
lp 1222 fs/ubifs/lprops.c dirty = lp->dirty;
lp 1225 fs/ubifs/lprops.c if (lp->free != free || lp->dirty != dirty)
lp 1228 fs/ubifs/lprops.c if (is_idx && !(lp->flags & LPROPS_INDEX)) {
lp 1239 fs/ubifs/lprops.c if (!is_idx && (lp->flags & LPROPS_INDEX)) {
lp 1250 fs/ubifs/lprops.c if (!(lp->flags & LPROPS_INDEX))
lp 1255 fs/ubifs/lprops.c if (!(lp->flags & LPROPS_INDEX)) {
lp 1271 fs/ubifs/lprops.c lnum, lp->free, lp->dirty, lp->flags, free, dirty);
lp 2069 fs/ubifs/lpt.c struct ubifs_lprops *lp, *lprops = &pnode->lprops[i];
lp 2144 fs/ubifs/lpt.c list_for_each_entry(lp, list, list)
lp 2145 fs/ubifs/lpt.c if (lprops == lp) {
lp 1085 fs/ubifs/recovery.c struct ubifs_lprops lp;
lp 1101 fs/ubifs/recovery.c err = ubifs_find_dirty_leb(c, &lp, wbuf->offs, 2);
lp 1107 fs/ubifs/recovery.c ubifs_assert(!(lp.flags & LPROPS_INDEX));
lp 1108 fs/ubifs/recovery.c lnum = lp.lnum;
lp 1109 fs/ubifs/recovery.c if (lp.free + lp.dirty == c->leb_size) {
lp 1111 fs/ubifs/recovery.c if (lp.free != c->leb_size) {
lp 1130 fs/ubifs/recovery.c if (lp.free + lp.dirty < wbuf->offs) {
lp 1152 fs/ubifs/recovery.c err = ubifs_garbage_collect_leb(c, &lp);
lp 108 fs/ubifs/replay.c const struct ubifs_lprops *lp;
lp 113 fs/ubifs/replay.c lp = ubifs_lpt_lookup_dirty(c, r->lnum);
lp 114 fs/ubifs/replay.c if (IS_ERR(lp)) {
lp 115 fs/ubifs/replay.c err = PTR_ERR(lp);
lp 119 fs/ubifs/replay.c dirty = lp->dirty;
lp 120 fs/ubifs/replay.c if (r->offs == 0 && (lp->free != c->leb_size || lp->dirty != 0)) {
lp 140 fs/ubifs/replay.c lp->free, lp->dirty);
lp 142 fs/ubifs/replay.c lp->free, lp->dirty);
lp 143 fs/ubifs/replay.c dirty -= c->leb_size - lp->free;
lp 153 fs/ubifs/replay.c "replay: %d free %d dirty", r->lnum, lp->free,
lp 154 fs/ubifs/replay.c lp->dirty, r->free, r->dirty);
lp 156 fs/ubifs/replay.c lp = ubifs_change_lp(c, lp, r->free, dirty + r->dirty,
lp 157 fs/ubifs/replay.c lp->flags | LPROPS_TAKEN, 0);
lp 158 fs/ubifs/replay.c if (IS_ERR(lp)) {
lp 159 fs/ubifs/replay.c err = PTR_ERR(lp);
lp 976 fs/ubifs/replay.c const struct ubifs_lprops *lp;
lp 981 fs/ubifs/replay.c lp = ubifs_lpt_lookup_dirty(c, c->ihead_lnum);
lp 982 fs/ubifs/replay.c if (IS_ERR(lp)) {
lp 983 fs/ubifs/replay.c err = PTR_ERR(lp);
lp 987 fs/ubifs/replay.c free = lp->free;
lp 989 fs/ubifs/replay.c lp = ubifs_change_lp(c, lp, LPROPS_NC, LPROPS_NC,
lp 990 fs/ubifs/replay.c lp->flags | LPROPS_TAKEN, 0);
lp 991 fs/ubifs/replay.c if (IS_ERR(lp)) {
lp 992 fs/ubifs/replay.c err = PTR_ERR(lp);
lp 297 fs/ubifs/tnc_commit.c struct ubifs_lprops lp;
lp 300 fs/ubifs/tnc_commit.c err = ubifs_read_one_lp(c, lnum, &lp);
lp 303 fs/ubifs/tnc_commit.c if (lp.free == c->leb_size) {
lp 436 fs/xfs/quota/xfs_dquot_item.c xfs_dq_logitem_t *lp;
lp 437 fs/xfs/quota/xfs_dquot_item.c lp = &dqp->q_logitem;
lp 439 fs/xfs/quota/xfs_dquot_item.c lp->qli_item.li_type = XFS_LI_DQUOT;
lp 440 fs/xfs/quota/xfs_dquot_item.c lp->qli_item.li_ops = &xfs_dquot_item_ops;
lp 441 fs/xfs/quota/xfs_dquot_item.c lp->qli_item.li_mountp = dqp->q_mount;
lp 442 fs/xfs/quota/xfs_dquot_item.c lp->qli_dquot = dqp;
lp 443 fs/xfs/quota/xfs_dquot_item.c lp->qli_format.qlf_type = XFS_LI_DQUOT;
lp 444 fs/xfs/quota/xfs_dquot_item.c lp->qli_format.qlf_id = be32_to_cpu(dqp->q_core.d_id);
lp 445 fs/xfs/quota/xfs_dquot_item.c lp->qli_format.qlf_blkno = dqp->q_blkno;
lp 446 fs/xfs/quota/xfs_dquot_item.c lp->qli_format.qlf_len = 1;
lp 454 fs/xfs/quota/xfs_dquot_item.c lp->qli_format.qlf_boffset = (__uint32_t)dqp->q_bufoffset;
lp 63 fs/xfs/quota/xfs_trans_dquot.c xfs_dq_logitem_t *lp;
lp 68 fs/xfs/quota/xfs_trans_dquot.c lp = &dqp->q_logitem;
lp 73 fs/xfs/quota/xfs_trans_dquot.c (void) xfs_trans_add_item(tp, (xfs_log_item_t*)(lp));
lp 2844 fs/xfs/xfs_attr_leaf.c xfs_attr_inactive_list_t *list, *lp;
lp 2881 fs/xfs/xfs_attr_leaf.c lp = list;
lp 2888 fs/xfs/xfs_attr_leaf.c lp->valueblk = be32_to_cpu(name_rmt->valueblk);
lp 2889 fs/xfs/xfs_attr_leaf.c lp->valuelen = XFS_B_TO_FSB(dp->i_mount,
lp 2891 fs/xfs/xfs_attr_leaf.c lp++;
lp 2901 fs/xfs/xfs_attr_leaf.c for (lp = list, i = 0; i < count; i++, lp++) {
lp 2903 fs/xfs/xfs_attr_leaf.c lp->valueblk, lp->valuelen);
lp 98 fs/xfs/xfs_dir2_leaf.h ((char *)(lp) + (mp)->m_dirblksize -
lp 2885 fs/xfs/xfs_log_recover.c xfs_caddr_t lp;
lp 2894 fs/xfs/xfs_log_recover.c lp = dp + be32_to_cpu(rhead->h_len);
lp 2901 fs/xfs/xfs_log_recover.c while ((dp < lp) && num_logops) {
lp 2902 fs/xfs/xfs_log_recover.c ASSERT(dp + sizeof(xlog_op_header_t) <= lp);
lp 2920 fs/xfs/xfs_log_recover.c if (dp + be32_to_cpu(ohead->oh_len) > lp) {
lp 1748 fs/xfs/xfs_vnodeops.c xfs_log_item_t *lp;
lp 1770 fs/xfs/xfs_vnodeops.c lp = (xfs_log_item_t *)ips[j]->i_itemp;
lp 1771 fs/xfs/xfs_vnodeops.c if (lp && (lp->li_flags & XFS_LI_IN_AIL)) {
lp 1855 fs/xfs/xfs_vnodeops.c xfs_log_item_t *lp;
lp 1875 fs/xfs/xfs_vnodeops.c lp = (xfs_log_item_t *)ip0->i_itemp;
lp 1876 fs/xfs/xfs_vnodeops.c if (lp && (lp->li_flags & XFS_LI_IN_AIL)) {
lp 323 include/linux/arcdevice.h #define ARCRESET(x) (lp->hw.reset(dev, (x)))
lp 324 include/linux/arcdevice.h #define ACOMMAND(x) (lp->hw.command(dev, (x)))
lp 325 include/linux/arcdevice.h #define ASTATUS() (lp->hw.status(dev))
lp 326 include/linux/arcdevice.h #define AINTMASK(x) (lp->hw.intmask(dev, (x)))
lp 85 include/linux/com20020.h lp->config = (lp->config & ~0x03) | (x); \
lp 99 include/linux/com20020.h #define ARCRESET { outb(lp->config | 0x80, _CONFIG); \
lp 101 include/linux/com20020.h outb(lp->config , _CONFIG); \
lp 113 include/linux/com20020.h #define SETCONF outb(lp->config, _CONFIG)
lp 225 include/linux/isdn_ppp.h struct isdn_net_local_s *lp;
lp 1428 include/linux/security.h struct sched_param *lp);
lp 2256 include/linux/security.h return cap_task_setscheduler(p, policy, lp);
lp 137 kernel/power/snapshot.c struct linked_page *lp = list->next;
lp 140 kernel/power/snapshot.c list = lp;
lp 180 kernel/power/snapshot.c struct linked_page *lp;
lp 182 kernel/power/snapshot.c lp = get_image_page(ca->gfp_mask, ca->safe_needed);
lp 183 kernel/power/snapshot.c if (!lp)
lp 186 kernel/power/snapshot.c lp->next = ca->chain;
lp 187 kernel/power/snapshot.c ca->chain = lp;
lp 1708 kernel/power/snapshot.c struct linked_page *sp_list, *lp;
lp 1742 kernel/power/snapshot.c lp = get_image_page(GFP_ATOMIC, PG_SAFE);
lp 1743 kernel/power/snapshot.c if (!lp) {
lp 1747 kernel/power/snapshot.c lp->next = sp_list;
lp 1748 kernel/power/snapshot.c sp_list = lp;
lp 1755 kernel/power/snapshot.c lp = (struct linked_page *)get_zeroed_page(GFP_ATOMIC);
lp 1756 kernel/power/snapshot.c if (!lp) {
lp 1760 kernel/power/snapshot.c if (!swsusp_page_is_free(virt_to_page(lp))) {
lp 1762 kernel/power/snapshot.c lp->next = safe_pages_list;
lp 1763 kernel/power/snapshot.c safe_pages_list = lp;
lp 1766 kernel/power/snapshot.c swsusp_set_page_forbidden(virt_to_page(lp));
lp 1767 kernel/power/snapshot.c swsusp_set_page_free(virt_to_page(lp));
lp 1772 kernel/power/snapshot.c lp = sp_list->next;
lp 1774 kernel/power/snapshot.c sp_list = lp;
lp 5356 kernel/sched.c struct sched_param lp;
lp 5373 kernel/sched.c lp.sched_priority = p->rt_priority;
lp 5379 kernel/sched.c retval = copy_to_user(param, &lp, sizeof(*param)) ? -EFAULT : 0;
lp 232 net/decnet/dn_neigh.c struct dn_long_packet *lp;
lp 251 net/decnet/dn_neigh.c lp = (struct dn_long_packet *)(data+3);
lp 256 net/decnet/dn_neigh.c lp->msgflg = DN_RT_PKT_LONG|(cb->rt_flags&(DN_RT_F_IE|DN_RT_F_RQR|DN_RT_F_RTS));
lp 257 net/decnet/dn_neigh.c lp->d_area = lp->d_subarea = 0;
lp 258 net/decnet/dn_neigh.c dn_dn2eth(lp->d_id, cb->dst);
lp 259 net/decnet/dn_neigh.c lp->s_area = lp->s_subarea = 0;
lp 260 net/decnet/dn_neigh.c dn_dn2eth(lp->s_id, cb->src);
lp 261 net/decnet/dn_neigh.c lp->nl2 = 0;
lp 262 net/decnet/dn_neigh.c lp->visit_ct = cb->hops & 0x3f;
lp 263 net/decnet/dn_neigh.c lp->s_class = 0;
lp 264 net/decnet/dn_neigh.c lp->pt = 0;
lp 680 net/ipv4/netfilter/nf_nat_snmp_basic.c unsigned long *lp, *id;
lp 773 net/ipv4/netfilter/nf_nat_snmp_basic.c if (!asn1_oid_decode(ctx, end, (unsigned long **)&lp, &len)) {
lp 780 net/ipv4/netfilter/nf_nat_snmp_basic.c kfree(lp);
lp 786 net/ipv4/netfilter/nf_nat_snmp_basic.c memcpy((*obj)->syntax.ul, lp, len);
lp 787 net/ipv4/netfilter/nf_nat_snmp_basic.c kfree(lp);
lp 97 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 99 net/ipv4/tcp_lp.c lp->flag = 0;
lp 100 net/ipv4/tcp_lp.c lp->sowd = 0;
lp 101 net/ipv4/tcp_lp.c lp->owd_min = 0xffffffff;
lp 102 net/ipv4/tcp_lp.c lp->owd_max = 0;
lp 103 net/ipv4/tcp_lp.c lp->owd_max_rsv = 0;
lp 104 net/ipv4/tcp_lp.c lp->remote_hz = 0;
lp 105 net/ipv4/tcp_lp.c lp->remote_ref_time = 0;
lp 106 net/ipv4/tcp_lp.c lp->local_ref_time = 0;
lp 107 net/ipv4/tcp_lp.c lp->last_drop = 0;
lp 108 net/ipv4/tcp_lp.c lp->inference = 0;
lp 120 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 122 net/ipv4/tcp_lp.c if (!(lp->flag & LP_WITHIN_INF))
lp 136 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 137 net/ipv4/tcp_lp.c s64 rhz = lp->remote_hz << 6; /* remote HZ << 6 */
lp 142 net/ipv4/tcp_lp.c if (lp->remote_ref_time == 0 || lp->local_ref_time == 0)
lp 146 net/ipv4/tcp_lp.c if (tp->rx_opt.rcv_tsval == lp->remote_ref_time
lp 147 net/ipv4/tcp_lp.c || tp->rx_opt.rcv_tsecr == lp->local_ref_time)
lp 151 net/ipv4/tcp_lp.c lp->remote_ref_time) / (tp->rx_opt.rcv_tsecr -
lp 152 net/ipv4/tcp_lp.c lp->local_ref_time);
lp 165 net/ipv4/tcp_lp.c lp->flag |= LP_VALID_RHZ;
lp 167 net/ipv4/tcp_lp.c lp->flag &= ~LP_VALID_RHZ;
lp 170 net/ipv4/tcp_lp.c lp->remote_ref_time = tp->rx_opt.rcv_tsval;
lp 171 net/ipv4/tcp_lp.c lp->local_ref_time = tp->rx_opt.rcv_tsecr;
lp 189 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 192 net/ipv4/tcp_lp.c lp->remote_hz = tcp_lp_remote_hz_estimator(sk);
lp 194 net/ipv4/tcp_lp.c if (lp->flag & LP_VALID_RHZ) {
lp 196 net/ipv4/tcp_lp.c tp->rx_opt.rcv_tsval * (LP_RESOL / lp->remote_hz) -
lp 203 net/ipv4/tcp_lp.c lp->flag |= LP_VALID_OWD;
lp 205 net/ipv4/tcp_lp.c lp->flag &= ~LP_VALID_OWD;
lp 222 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 226 net/ipv4/tcp_lp.c if (!(lp->flag & LP_VALID_RHZ) || !(lp->flag & LP_VALID_OWD))
lp 230 net/ipv4/tcp_lp.c if (mowd < lp->owd_min)
lp 231 net/ipv4/tcp_lp.c lp->owd_min = mowd;
lp 235 net/ipv4/tcp_lp.c if (mowd > lp->owd_max) {
lp 236 net/ipv4/tcp_lp.c if (mowd > lp->owd_max_rsv) {
lp 237 net/ipv4/tcp_lp.c if (lp->owd_max_rsv == 0)
lp 238 net/ipv4/tcp_lp.c lp->owd_max = mowd;
lp 240 net/ipv4/tcp_lp.c lp->owd_max = lp->owd_max_rsv;
lp 241 net/ipv4/tcp_lp.c lp->owd_max_rsv = mowd;
lp 243 net/ipv4/tcp_lp.c lp->owd_max = mowd;
lp 247 net/ipv4/tcp_lp.c if (lp->sowd != 0) {
lp 248 net/ipv4/tcp_lp.c mowd -= lp->sowd >> 3; /* m is now error in owd est */
lp 249 net/ipv4/tcp_lp.c lp->sowd += mowd; /* owd = 7/8 owd + 1/8 new */
lp 251 net/ipv4/tcp_lp.c lp->sowd = mowd << 3; /* take the measured time be owd */
lp 266 net/ipv4/tcp_lp.c struct lp *lp = inet_csk_ca(sk);
lp 273 net/ipv4/tcp_lp.c lp->inference = 3 * (tcp_time_stamp - tp->rx_opt.rcv_tsecr);
lp 276 net/ipv4/tcp_lp.c if (lp->last_drop && (tcp_time_stamp - lp->last_drop < lp->inference))
lp 277 net/ipv4/tcp_lp.c lp->flag |= LP_WITHIN_INF;
lp 279 net/ipv4/tcp_lp.c lp->flag &= ~LP_WITHIN_INF;
lp 282 net/ipv4/tcp_lp.c if (lp->sowd >> 3 <
lp 283 net/ipv4/tcp_lp.c lp->owd_min + 15 * (lp->owd_max - lp->owd_min) / 100)
lp 284 net/ipv4/tcp_lp.c lp->flag |= LP_WITHIN_THR;
lp 286 net/ipv4/tcp_lp.c lp->flag &= ~LP_WITHIN_THR;
lp 288 net/ipv4/tcp_lp.c pr_debug("TCP-LP: %05o|%5u|%5u|%15u|%15u|%15u\n", lp->flag,
lp 289 net/ipv4/tcp_lp.c tp->snd_cwnd, lp->remote_hz, lp->owd_min, lp->owd_max,
lp 290 net/ipv4/tcp_lp.c lp->sowd >> 3);
lp 292 net/ipv4/tcp_lp.c if (lp->flag & LP_WITHIN_THR)
lp 298 net/ipv4/tcp_lp.c lp->owd_min = lp->sowd >> 3;
lp 299 net/ipv4/tcp_lp.c lp->owd_max = lp->sowd >> 2;
lp 300 net/ipv4/tcp_lp.c lp->owd_max_rsv = lp->sowd >> 2;
lp 304 net/ipv4/tcp_lp.c if (lp->flag & LP_WITHIN_INF)
lp 313 net/ipv4/tcp_lp.c lp->last_drop = tcp_time_stamp;
lp 330 net/ipv4/tcp_lp.c BUILD_BUG_ON(sizeof(struct lp) > ICSK_CA_PRIV_SIZE);
lp 57 net/ipv6/netfilter/ip6t_hbh.c const u_int8_t *lp = NULL;
lp 124 net/ipv6/netfilter/ip6t_hbh.c lp = skb_header_pointer(skb, ptr + 1,
lp 127 net/ipv6/netfilter/ip6t_hbh.c if (lp == NULL)
lp 131 net/ipv6/netfilter/ip6t_hbh.c if (spec_len != 0x00FF && spec_len != *lp) {
lp 132 net/ipv6/netfilter/ip6t_hbh.c pr_debug("Lbad %02X %04X\n", *lp,
lp 137 net/ipv6/netfilter/ip6t_hbh.c optlen = *lp + 2;
lp 607 net/sunrpc/cache.c struct list_head *lp;
lp 613 net/sunrpc/cache.c lp = cache_defer_hash[hash].next;
lp 614 net/sunrpc/cache.c if (lp) {
lp 615 net/sunrpc/cache.c while (lp != &cache_defer_hash[hash]) {
lp 616 net/sunrpc/cache.c dreq = list_entry(lp, struct cache_deferred_req, hash);
lp 617 net/sunrpc/cache.c lp = lp->next;
lp 960 net/sunrpc/cache.c int len = *lp;
lp 989 net/sunrpc/cache.c *lp = len;
lp 996 net/sunrpc/cache.c int len = *lp;
lp 1018 net/sunrpc/cache.c *lp = len;
lp 682 security/security.c return security_ops->task_setscheduler(p, policy, lp);
lp 3316 security/selinux/hooks.c rc = secondary_ops->task_setscheduler(p, policy, lp);
lp 1328 security/selinux/ss/policydb.c memset(lp, 0, sizeof(*lp));
lp 1335 security/selinux/ss/policydb.c lp->sens = le32_to_cpu(buf[0]);
lp 1337 security/selinux/ss/policydb.c if (ebitmap_read(&lp->cat, fp)) {
lp 1109 security/smack/smack_lsm.c rc = cap_task_setscheduler(p, policy, lp);