struct audit_chunk *audit_tree_lookup(const struct inode *inode)
{
struct list_head *list = chunk_hash(inode);
- struct list_head *pos;
+ struct audit_chunk *p;
- list_for_each_rcu(pos, list) {
- struct audit_chunk *p = container_of(pos, struct audit_chunk, hash);
+ list_for_each_entry_rcu(p, list, hash) {
if (p->watch.inode == inode) {
get_inotify_watch(&p->watch);
return p;
goto skip_it;
root_mnt = collect_mounts(nd.path.mnt, nd.path.dentry);
- path_release(&nd);
+ path_put(&nd.path);
if (!root_mnt)
goto skip_it;
if (err)
goto Err;
mnt = collect_mounts(nd.path.mnt, nd.path.dentry);
- path_release(&nd);
+ path_put(&nd.path);
if (!mnt) {
err = -ENOMEM;
goto Err;
if (err)
return err;
tagged = collect_mounts(nd.path.mnt, nd.path.dentry);
- path_release(&nd);
+ path_put(&nd.path);
if (!tagged)
return -ENOMEM;
}
mnt = mntget(nd.path.mnt);
dentry = dget(nd.path.dentry);
- path_release(&nd);
+ path_put(&nd.path);
if (dentry == tagged->mnt_root && dentry == mnt->mnt_root)
follow_up(&mnt, &dentry);
spin_lock(&vfsmount_lock);
if (!is_under(mnt, dentry, &nd)) {
spin_unlock(&vfsmount_lock);
- path_release(&nd);
+ path_put(&nd.path);
put_tree(tree);
mutex_lock(&audit_filter_mutex);
continue;
}
spin_unlock(&vfsmount_lock);
- path_release(&nd);
+ path_put(&nd.path);
list_for_each_entry(p, &list, mnt_list) {
failed = tag_chunk(p->mnt_root->d_inode, tree);