• Home
  • Raw
  • Download

Lines Matching full:ui

288 		struct ubifs_inode *ui = ubifs_inode(inode);  in write_begin_slow()  local
295 mutex_lock(&ui->ui_mutex); in write_begin_slow()
296 if (ui->dirty) in write_begin_slow()
301 ubifs_release_dirty_inode_budget(c, ui); in write_begin_slow()
312 * @ui: UBIFS inode object the page belongs to
318 * function leaves the @ui->ui_mutex locked in case of appending. Returns zero
322 struct ubifs_inode *ui, int appending) in allocate_budget() argument
334 mutex_lock(&ui->ui_mutex); in allocate_budget()
335 if (ui->dirty) in allocate_budget()
340 * but @ui->ui_mutex hast to be left locked because we in allocate_budget()
371 mutex_lock(&ui->ui_mutex); in allocate_budget()
372 if (!ui->dirty) in allocate_budget()
423 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_write_begin() local
464 err = allocate_budget(c, page, ui, appending); in ubifs_write_begin()
481 ubifs_assert(c, mutex_is_locked(&ui->ui_mutex)); in ubifs_write_begin()
482 mutex_unlock(&ui->ui_mutex); in ubifs_write_begin()
493 * with @ui->ui_mutex locked if we are appending pages, and unlocked in ubifs_write_begin()
505 * @ui: UBIFS inode object the page belongs to
509 * @ui->ui_mutex in case of appending.
512 struct ubifs_inode *ui, int appending) in cancel_budget() argument
515 if (!ui->dirty) in cancel_budget()
516 ubifs_release_dirty_inode_budget(c, ui); in cancel_budget()
517 mutex_unlock(&ui->ui_mutex); in cancel_budget()
532 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_write_end() local
552 cancel_budget(c, page, ui, appending); in ubifs_write_end()
574 ui->ui_size = end_pos; in ubifs_write_end()
581 ubifs_assert(c, mutex_is_locked(&ui->ui_mutex)); in ubifs_write_end()
582 mutex_unlock(&ui->ui_mutex); in ubifs_write_end()
717 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_do_bulk_read() local
729 ui->read_in_a_row = 1; in ubifs_do_bulk_read()
730 ui->bulk_read = 0; in ubifs_do_bulk_read()
796 ui->last_page_read = offset + page_idx - 1; in ubifs_do_bulk_read()
808 ui->read_in_a_row = ui->bulk_read = 0; in ubifs_do_bulk_read()
825 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_bulk_read() local
826 pgoff_t index = page->index, last_page_read = ui->last_page_read; in ubifs_bulk_read()
830 ui->last_page_read = index; in ubifs_bulk_read()
835 * Bulk-read is protected by @ui->ui_mutex, but it is an optimization, in ubifs_bulk_read()
838 if (!mutex_trylock(&ui->ui_mutex)) in ubifs_bulk_read()
843 ui->read_in_a_row = 1; in ubifs_bulk_read()
844 if (ui->bulk_read) in ubifs_bulk_read()
845 ui->bulk_read = 0; in ubifs_bulk_read()
849 if (!ui->bulk_read) { in ubifs_bulk_read()
850 ui->read_in_a_row += 1; in ubifs_bulk_read()
851 if (ui->read_in_a_row < 3) in ubifs_bulk_read()
854 ui->bulk_read = 1; in ubifs_bulk_read()
883 mutex_unlock(&ui->ui_mutex); in ubifs_bulk_read()
908 struct ubifs_inode *ui = ubifs_inode(inode); in do_writepage() local
909 spin_lock(&ui->ui_lock); in do_writepage()
910 ubifs_assert(c, page->index <= ui->synced_i_size >> PAGE_SHIFT); in do_writepage()
911 spin_unlock(&ui->ui_lock); in do_writepage()
991 * @ui->ui_isize "shadow" field which UBIFS uses instead of @inode->i_size
1005 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_writepage() local
1021 spin_lock(&ui->ui_lock); in ubifs_writepage()
1022 synced_i_size = ui->synced_i_size; in ubifs_writepage()
1023 spin_unlock(&ui->ui_lock); in ubifs_writepage()
1117 struct ubifs_inode *ui = ubifs_inode(inode); in do_truncation() local
1187 mutex_lock(&ui->ui_mutex); in do_truncation()
1188 ui->ui_size = inode->i_size; in do_truncation()
1194 mutex_unlock(&ui->ui_mutex); in do_truncation()
1221 struct ubifs_inode *ui = ubifs_inode(inode); in do_setattr() local
1223 .dirtied_ino_d = ALIGN(ui->data_len, 8) }; in do_setattr()
1234 mutex_lock(&ui->ui_mutex); in do_setattr()
1239 ui->ui_size = inode->i_size; in do_setattr()
1244 release = ui->dirty; in do_setattr()
1253 mutex_unlock(&ui->ui_mutex); in do_setattr()
1378 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_update_time() local
1381 .dirtied_ino_d = ALIGN(ui->data_len, 8) }; in ubifs_update_time()
1393 mutex_lock(&ui->ui_mutex); in ubifs_update_time()
1395 release = ui->dirty; in ubifs_update_time()
1397 mutex_unlock(&ui->ui_mutex); in ubifs_update_time()
1414 struct ubifs_inode *ui = ubifs_inode(inode); in update_mctime() local
1420 .dirtied_ino_d = ALIGN(ui->data_len, 8) }; in update_mctime()
1426 mutex_lock(&ui->ui_mutex); in update_mctime()
1428 release = ui->dirty; in update_mctime()
1430 mutex_unlock(&ui->ui_mutex); in update_mctime()
1562 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_vm_page_mkwrite() local
1564 mutex_lock(&ui->ui_mutex); in ubifs_vm_page_mkwrite()
1566 release = ui->dirty; in ubifs_vm_page_mkwrite()
1568 mutex_unlock(&ui->ui_mutex); in ubifs_vm_page_mkwrite()
1570 ubifs_release_dirty_inode_budget(c, ui); in ubifs_vm_page_mkwrite()
1607 struct ubifs_inode *ui = ubifs_inode(inode); in ubifs_get_link() local
1610 return ui->data; in ubifs_get_link()
1615 return fscrypt_get_symlink(inode, ui->data, ui->data_len, done); in ubifs_get_link()