Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/*
2 * linux/fs/hfs/inode.c
3 *
4 * Copyright (C) 1995-1997 Paul H. Hargrove
5 * (C) 2003 Ardis Technologies <roman@ardistech.com>
6 * This file may be distributed under the terms of the GNU General Public License.
7 *
8 * This file contains inode-related functions which do not depend on
9 * which scheme is being used to represent forks.
10 *
11 * Based on the minix file system code, (C) 1991, 1992 by Linus Torvalds
12 */
13
14#include <linux/pagemap.h>
15#include <linux/mpage.h>
16#include <linux/sched.h>
17#include <linux/cred.h>
18#include <linux/uio.h>
19#include <linux/xattr.h>
20#include <linux/blkdev.h>
21
22#include "hfs_fs.h"
23#include "btree.h"
24
25static const struct file_operations hfs_file_operations;
26static const struct inode_operations hfs_file_inode_operations;
27
28/*================ Variable-like macros ================*/
29
30#define HFS_VALID_MODE_BITS (S_IFREG | S_IFDIR | S_IRWXUGO)
31
32static int hfs_read_folio(struct file *file, struct folio *folio)
33{
34 return block_read_full_folio(folio, hfs_get_block);
35}
36
37static void hfs_write_failed(struct address_space *mapping, loff_t to)
38{
39 struct inode *inode = mapping->host;
40
41 if (to > inode->i_size) {
42 truncate_pagecache(inode, inode->i_size);
43 hfs_file_truncate(inode);
44 }
45}
46
47int hfs_write_begin(const struct kiocb *iocb, struct address_space *mapping,
48 loff_t pos, unsigned int len, struct folio **foliop,
49 void **fsdata)
50{
51 int ret;
52
53 ret = cont_write_begin(iocb, mapping, pos, len, foliop, fsdata,
54 hfs_get_block,
55 &HFS_I(mapping->host)->phys_size);
56 if (unlikely(ret))
57 hfs_write_failed(mapping, pos + len);
58
59 return ret;
60}
61
62static sector_t hfs_bmap(struct address_space *mapping, sector_t block)
63{
64 return generic_block_bmap(mapping, block, hfs_get_block);
65}
66
67static bool hfs_release_folio(struct folio *folio, gfp_t mask)
68{
69 struct inode *inode = folio->mapping->host;
70 struct super_block *sb = inode->i_sb;
71 struct hfs_btree *tree;
72 struct hfs_bnode *node;
73 u32 nidx;
74 int i;
75 bool res = true;
76
77 switch (inode->i_ino) {
78 case HFS_EXT_CNID:
79 tree = HFS_SB(sb)->ext_tree;
80 break;
81 case HFS_CAT_CNID:
82 tree = HFS_SB(sb)->cat_tree;
83 break;
84 default:
85 BUG();
86 return false;
87 }
88
89 if (!tree)
90 return false;
91
92 if (tree->node_size >= PAGE_SIZE) {
93 nidx = folio->index >> (tree->node_size_shift - PAGE_SHIFT);
94 spin_lock(&tree->hash_lock);
95 node = hfs_bnode_findhash(tree, nidx);
96 if (!node)
97 ;
98 else if (atomic_read(&node->refcnt))
99 res = false;
100 if (res && node) {
101 hfs_bnode_unhash(node);
102 hfs_bnode_free(node);
103 }
104 spin_unlock(&tree->hash_lock);
105 } else {
106 nidx = folio->index << (PAGE_SHIFT - tree->node_size_shift);
107 i = 1 << (PAGE_SHIFT - tree->node_size_shift);
108 spin_lock(&tree->hash_lock);
109 do {
110 node = hfs_bnode_findhash(tree, nidx++);
111 if (!node)
112 continue;
113 if (atomic_read(&node->refcnt)) {
114 res = false;
115 break;
116 }
117 hfs_bnode_unhash(node);
118 hfs_bnode_free(node);
119 } while (--i && nidx < tree->node_count);
120 spin_unlock(&tree->hash_lock);
121 }
122 return res ? try_to_free_buffers(folio) : false;
123}
124
125static ssize_t hfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
126{
127 struct file *file = iocb->ki_filp;
128 struct address_space *mapping = file->f_mapping;
129 struct inode *inode = mapping->host;
130 size_t count = iov_iter_count(iter);
131 ssize_t ret;
132
133 ret = blockdev_direct_IO(iocb, inode, iter, hfs_get_block);
134
135 /*
136 * In case of error extending write may have instantiated a few
137 * blocks outside i_size. Trim these off again.
138 */
139 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) {
140 loff_t isize = i_size_read(inode);
141 loff_t end = iocb->ki_pos + count;
142
143 if (end > isize)
144 hfs_write_failed(mapping, end);
145 }
146
147 return ret;
148}
149
150static int hfs_writepages(struct address_space *mapping,
151 struct writeback_control *wbc)
152{
153 return mpage_writepages(mapping, wbc, hfs_get_block);
154}
155
156const struct address_space_operations hfs_btree_aops = {
157 .dirty_folio = block_dirty_folio,
158 .invalidate_folio = block_invalidate_folio,
159 .read_folio = hfs_read_folio,
160 .writepages = hfs_writepages,
161 .write_begin = hfs_write_begin,
162 .write_end = generic_write_end,
163 .migrate_folio = buffer_migrate_folio,
164 .bmap = hfs_bmap,
165 .release_folio = hfs_release_folio,
166};
167
168const struct address_space_operations hfs_aops = {
169 .dirty_folio = block_dirty_folio,
170 .invalidate_folio = block_invalidate_folio,
171 .read_folio = hfs_read_folio,
172 .write_begin = hfs_write_begin,
173 .write_end = generic_write_end,
174 .bmap = hfs_bmap,
175 .direct_IO = hfs_direct_IO,
176 .writepages = hfs_writepages,
177 .migrate_folio = buffer_migrate_folio,
178};
179
180/*
181 * hfs_new_inode
182 */
183struct inode *hfs_new_inode(struct inode *dir, const struct qstr *name, umode_t mode)
184{
185 struct super_block *sb = dir->i_sb;
186 struct inode *inode = new_inode(sb);
187 s64 next_id;
188 s64 file_count;
189 s64 folder_count;
190
191 if (!inode)
192 return NULL;
193
194 mutex_init(&HFS_I(inode)->extents_lock);
195 INIT_LIST_HEAD(&HFS_I(inode)->open_dir_list);
196 spin_lock_init(&HFS_I(inode)->open_dir_lock);
197 hfs_cat_build_key(sb, (btree_key *)&HFS_I(inode)->cat_key, dir->i_ino, name);
198 next_id = atomic64_inc_return(&HFS_SB(sb)->next_id);
199 BUG_ON(next_id > U32_MAX);
200 inode->i_ino = (u32)next_id;
201 inode->i_mode = mode;
202 inode->i_uid = current_fsuid();
203 inode->i_gid = current_fsgid();
204 set_nlink(inode, 1);
205 simple_inode_init_ts(inode);
206 HFS_I(inode)->flags = 0;
207 HFS_I(inode)->rsrc_inode = NULL;
208 HFS_I(inode)->fs_blocks = 0;
209 HFS_I(inode)->tz_secondswest = sys_tz.tz_minuteswest * 60;
210 if (S_ISDIR(mode)) {
211 inode->i_size = 2;
212 folder_count = atomic64_inc_return(&HFS_SB(sb)->folder_count);
213 BUG_ON(folder_count > U32_MAX);
214 if (dir->i_ino == HFS_ROOT_CNID)
215 HFS_SB(sb)->root_dirs++;
216 inode->i_op = &hfs_dir_inode_operations;
217 inode->i_fop = &hfs_dir_operations;
218 inode->i_mode |= S_IRWXUGO;
219 inode->i_mode &= ~HFS_SB(inode->i_sb)->s_dir_umask;
220 } else if (S_ISREG(mode)) {
221 HFS_I(inode)->clump_blocks = HFS_SB(sb)->clumpablks;
222 file_count = atomic64_inc_return(&HFS_SB(sb)->file_count);
223 BUG_ON(file_count > U32_MAX);
224 if (dir->i_ino == HFS_ROOT_CNID)
225 HFS_SB(sb)->root_files++;
226 inode->i_op = &hfs_file_inode_operations;
227 inode->i_fop = &hfs_file_operations;
228 inode->i_mapping->a_ops = &hfs_aops;
229 inode->i_mode |= S_IRUGO|S_IXUGO;
230 if (mode & S_IWUSR)
231 inode->i_mode |= S_IWUGO;
232 inode->i_mode &= ~HFS_SB(inode->i_sb)->s_file_umask;
233 HFS_I(inode)->phys_size = 0;
234 HFS_I(inode)->alloc_blocks = 0;
235 HFS_I(inode)->first_blocks = 0;
236 HFS_I(inode)->cached_start = 0;
237 HFS_I(inode)->cached_blocks = 0;
238 memset(HFS_I(inode)->first_extents, 0, sizeof(hfs_extent_rec));
239 memset(HFS_I(inode)->cached_extents, 0, sizeof(hfs_extent_rec));
240 }
241 insert_inode_hash(inode);
242 mark_inode_dirty(inode);
243 set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
244 hfs_mark_mdb_dirty(sb);
245
246 return inode;
247}
248
249void hfs_delete_inode(struct inode *inode)
250{
251 struct super_block *sb = inode->i_sb;
252
253 hfs_dbg("ino %lu\n", inode->i_ino);
254 if (S_ISDIR(inode->i_mode)) {
255 BUG_ON(atomic64_read(&HFS_SB(sb)->folder_count) > U32_MAX);
256 atomic64_dec(&HFS_SB(sb)->folder_count);
257 if (HFS_I(inode)->cat_key.ParID == cpu_to_be32(HFS_ROOT_CNID))
258 HFS_SB(sb)->root_dirs--;
259 set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
260 hfs_mark_mdb_dirty(sb);
261 return;
262 }
263
264 BUG_ON(atomic64_read(&HFS_SB(sb)->file_count) > U32_MAX);
265 atomic64_dec(&HFS_SB(sb)->file_count);
266 if (HFS_I(inode)->cat_key.ParID == cpu_to_be32(HFS_ROOT_CNID))
267 HFS_SB(sb)->root_files--;
268 if (S_ISREG(inode->i_mode)) {
269 if (!inode->i_nlink) {
270 inode->i_size = 0;
271 hfs_file_truncate(inode);
272 }
273 }
274 set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
275 hfs_mark_mdb_dirty(sb);
276}
277
278void hfs_inode_read_fork(struct inode *inode, struct hfs_extent *ext,
279 __be32 __log_size, __be32 phys_size, u32 clump_size)
280{
281 struct super_block *sb = inode->i_sb;
282 u32 log_size = be32_to_cpu(__log_size);
283 u16 count;
284 int i;
285
286 memcpy(HFS_I(inode)->first_extents, ext, sizeof(hfs_extent_rec));
287 for (count = 0, i = 0; i < 3; i++)
288 count += be16_to_cpu(ext[i].count);
289 HFS_I(inode)->first_blocks = count;
290 HFS_I(inode)->cached_start = 0;
291 HFS_I(inode)->cached_blocks = 0;
292
293 inode->i_size = HFS_I(inode)->phys_size = log_size;
294 HFS_I(inode)->fs_blocks = (log_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
295 inode_set_bytes(inode, HFS_I(inode)->fs_blocks << sb->s_blocksize_bits);
296 HFS_I(inode)->alloc_blocks = be32_to_cpu(phys_size) /
297 HFS_SB(sb)->alloc_blksz;
298 HFS_I(inode)->clump_blocks = clump_size / HFS_SB(sb)->alloc_blksz;
299 if (!HFS_I(inode)->clump_blocks)
300 HFS_I(inode)->clump_blocks = HFS_SB(sb)->clumpablks;
301}
302
303struct hfs_iget_data {
304 struct hfs_cat_key *key;
305 hfs_cat_rec *rec;
306};
307
308static int hfs_test_inode(struct inode *inode, void *data)
309{
310 struct hfs_iget_data *idata = data;
311 hfs_cat_rec *rec;
312
313 rec = idata->rec;
314 switch (rec->type) {
315 case HFS_CDR_DIR:
316 return inode->i_ino == be32_to_cpu(rec->dir.DirID);
317 case HFS_CDR_FIL:
318 return inode->i_ino == be32_to_cpu(rec->file.FlNum);
319 default:
320 BUG();
321 return 1;
322 }
323}
324
325/*
326 * hfs_read_inode
327 */
328static int hfs_read_inode(struct inode *inode, void *data)
329{
330 struct hfs_iget_data *idata = data;
331 struct hfs_sb_info *hsb = HFS_SB(inode->i_sb);
332 hfs_cat_rec *rec;
333
334 HFS_I(inode)->flags = 0;
335 HFS_I(inode)->rsrc_inode = NULL;
336 mutex_init(&HFS_I(inode)->extents_lock);
337 INIT_LIST_HEAD(&HFS_I(inode)->open_dir_list);
338 spin_lock_init(&HFS_I(inode)->open_dir_lock);
339
340 /* Initialize the inode */
341 inode->i_uid = hsb->s_uid;
342 inode->i_gid = hsb->s_gid;
343 set_nlink(inode, 1);
344
345 if (idata->key)
346 HFS_I(inode)->cat_key = *idata->key;
347 else
348 HFS_I(inode)->flags |= HFS_FLG_RSRC;
349 HFS_I(inode)->tz_secondswest = sys_tz.tz_minuteswest * 60;
350
351 rec = idata->rec;
352 switch (rec->type) {
353 case HFS_CDR_FIL:
354 if (!HFS_IS_RSRC(inode)) {
355 hfs_inode_read_fork(inode, rec->file.ExtRec, rec->file.LgLen,
356 rec->file.PyLen, be16_to_cpu(rec->file.ClpSize));
357 } else {
358 hfs_inode_read_fork(inode, rec->file.RExtRec, rec->file.RLgLen,
359 rec->file.RPyLen, be16_to_cpu(rec->file.ClpSize));
360 }
361
362 inode->i_ino = be32_to_cpu(rec->file.FlNum);
363 inode->i_mode = S_IRUGO | S_IXUGO;
364 if (!(rec->file.Flags & HFS_FIL_LOCK))
365 inode->i_mode |= S_IWUGO;
366 inode->i_mode &= ~hsb->s_file_umask;
367 inode->i_mode |= S_IFREG;
368 inode_set_mtime_to_ts(inode,
369 inode_set_atime_to_ts(inode, inode_set_ctime_to_ts(inode, hfs_m_to_utime(rec->file.MdDat))));
370 inode->i_op = &hfs_file_inode_operations;
371 inode->i_fop = &hfs_file_operations;
372 inode->i_mapping->a_ops = &hfs_aops;
373 break;
374 case HFS_CDR_DIR:
375 inode->i_ino = be32_to_cpu(rec->dir.DirID);
376 inode->i_size = be16_to_cpu(rec->dir.Val) + 2;
377 HFS_I(inode)->fs_blocks = 0;
378 inode->i_mode = S_IFDIR | (S_IRWXUGO & ~hsb->s_dir_umask);
379 inode_set_mtime_to_ts(inode,
380 inode_set_atime_to_ts(inode, inode_set_ctime_to_ts(inode, hfs_m_to_utime(rec->dir.MdDat))));
381 inode->i_op = &hfs_dir_inode_operations;
382 inode->i_fop = &hfs_dir_operations;
383 break;
384 default:
385 make_bad_inode(inode);
386 }
387 return 0;
388}
389
390/*
391 * __hfs_iget()
392 *
393 * Given the MDB for a HFS filesystem, a 'key' and an 'entry' in
394 * the catalog B-tree and the 'type' of the desired file return the
395 * inode for that file/directory or NULL. Note that 'type' indicates
396 * whether we want the actual file or directory, or the corresponding
397 * metadata (AppleDouble header file or CAP metadata file).
398 */
399struct inode *hfs_iget(struct super_block *sb, struct hfs_cat_key *key, hfs_cat_rec *rec)
400{
401 struct hfs_iget_data data = { key, rec };
402 struct inode *inode;
403 u32 cnid;
404
405 switch (rec->type) {
406 case HFS_CDR_DIR:
407 cnid = be32_to_cpu(rec->dir.DirID);
408 break;
409 case HFS_CDR_FIL:
410 cnid = be32_to_cpu(rec->file.FlNum);
411 break;
412 default:
413 return NULL;
414 }
415 inode = iget5_locked(sb, cnid, hfs_test_inode, hfs_read_inode, &data);
416 if (inode && (inode_state_read_once(inode) & I_NEW))
417 unlock_new_inode(inode);
418 return inode;
419}
420
421void hfs_inode_write_fork(struct inode *inode, struct hfs_extent *ext,
422 __be32 *log_size, __be32 *phys_size)
423{
424 memcpy(ext, HFS_I(inode)->first_extents, sizeof(hfs_extent_rec));
425
426 if (log_size)
427 *log_size = cpu_to_be32(inode->i_size);
428 if (phys_size)
429 *phys_size = cpu_to_be32(HFS_I(inode)->alloc_blocks *
430 HFS_SB(inode->i_sb)->alloc_blksz);
431}
432
433int hfs_write_inode(struct inode *inode, struct writeback_control *wbc)
434{
435 struct inode *main_inode = inode;
436 struct hfs_find_data fd;
437 hfs_cat_rec rec;
438 int res;
439
440 hfs_dbg("ino %lu\n", inode->i_ino);
441 res = hfs_ext_write_extent(inode);
442 if (res)
443 return res;
444
445 if (inode->i_ino < HFS_FIRSTUSER_CNID) {
446 switch (inode->i_ino) {
447 case HFS_ROOT_CNID:
448 break;
449 case HFS_EXT_CNID:
450 hfs_btree_write(HFS_SB(inode->i_sb)->ext_tree);
451 return 0;
452 case HFS_CAT_CNID:
453 hfs_btree_write(HFS_SB(inode->i_sb)->cat_tree);
454 return 0;
455 default:
456 BUG();
457 return -EIO;
458 }
459 }
460
461 if (HFS_IS_RSRC(inode))
462 main_inode = HFS_I(inode)->rsrc_inode;
463
464 if (!main_inode->i_nlink)
465 return 0;
466
467 if (hfs_find_init(HFS_SB(main_inode->i_sb)->cat_tree, &fd))
468 /* panic? */
469 return -EIO;
470
471 res = -EIO;
472 if (HFS_I(main_inode)->cat_key.CName.len > HFS_NAMELEN)
473 goto out;
474 fd.search_key->cat = HFS_I(main_inode)->cat_key;
475 if (hfs_brec_find(&fd))
476 goto out;
477
478 if (S_ISDIR(main_inode->i_mode)) {
479 if (fd.entrylength < sizeof(struct hfs_cat_dir))
480 goto out;
481 hfs_bnode_read(fd.bnode, &rec, fd.entryoffset,
482 sizeof(struct hfs_cat_dir));
483 if (rec.type != HFS_CDR_DIR ||
484 be32_to_cpu(rec.dir.DirID) != inode->i_ino) {
485 }
486
487 rec.dir.MdDat = hfs_u_to_mtime(inode_get_mtime(inode));
488 rec.dir.Val = cpu_to_be16(inode->i_size - 2);
489
490 hfs_bnode_write(fd.bnode, &rec, fd.entryoffset,
491 sizeof(struct hfs_cat_dir));
492 } else if (HFS_IS_RSRC(inode)) {
493 if (fd.entrylength < sizeof(struct hfs_cat_file))
494 goto out;
495 hfs_bnode_read(fd.bnode, &rec, fd.entryoffset,
496 sizeof(struct hfs_cat_file));
497 hfs_inode_write_fork(inode, rec.file.RExtRec,
498 &rec.file.RLgLen, &rec.file.RPyLen);
499 hfs_bnode_write(fd.bnode, &rec, fd.entryoffset,
500 sizeof(struct hfs_cat_file));
501 } else {
502 if (fd.entrylength < sizeof(struct hfs_cat_file))
503 goto out;
504 hfs_bnode_read(fd.bnode, &rec, fd.entryoffset,
505 sizeof(struct hfs_cat_file));
506 if (rec.type != HFS_CDR_FIL ||
507 be32_to_cpu(rec.file.FlNum) != inode->i_ino) {
508 }
509
510 if (inode->i_mode & S_IWUSR)
511 rec.file.Flags &= ~HFS_FIL_LOCK;
512 else
513 rec.file.Flags |= HFS_FIL_LOCK;
514 hfs_inode_write_fork(inode, rec.file.ExtRec, &rec.file.LgLen, &rec.file.PyLen);
515 rec.file.MdDat = hfs_u_to_mtime(inode_get_mtime(inode));
516
517 hfs_bnode_write(fd.bnode, &rec, fd.entryoffset,
518 sizeof(struct hfs_cat_file));
519 }
520 res = 0;
521out:
522 hfs_find_exit(&fd);
523 return res;
524}
525
526static struct dentry *hfs_file_lookup(struct inode *dir, struct dentry *dentry,
527 unsigned int flags)
528{
529 struct inode *inode = NULL;
530 hfs_cat_rec rec;
531 struct hfs_find_data fd;
532 int res;
533
534 if (HFS_IS_RSRC(dir) || strcmp(dentry->d_name.name, "rsrc"))
535 goto out;
536
537 inode = HFS_I(dir)->rsrc_inode;
538 if (inode)
539 goto out;
540
541 inode = new_inode(dir->i_sb);
542 if (!inode)
543 return ERR_PTR(-ENOMEM);
544
545 res = hfs_find_init(HFS_SB(dir->i_sb)->cat_tree, &fd);
546 if (res) {
547 iput(inode);
548 return ERR_PTR(res);
549 }
550 fd.search_key->cat = HFS_I(dir)->cat_key;
551 res = hfs_brec_read(&fd, &rec, sizeof(rec));
552 if (!res) {
553 struct hfs_iget_data idata = { NULL, &rec };
554 hfs_read_inode(inode, &idata);
555 }
556 hfs_find_exit(&fd);
557 if (res) {
558 iput(inode);
559 return ERR_PTR(res);
560 }
561 HFS_I(inode)->rsrc_inode = dir;
562 HFS_I(dir)->rsrc_inode = inode;
563 igrab(dir);
564 inode_fake_hash(inode);
565 mark_inode_dirty(inode);
566 dont_mount(dentry);
567out:
568 return d_splice_alias(inode, dentry);
569}
570
571void hfs_evict_inode(struct inode *inode)
572{
573 truncate_inode_pages_final(&inode->i_data);
574 clear_inode(inode);
575 if (HFS_IS_RSRC(inode) && HFS_I(inode)->rsrc_inode) {
576 HFS_I(HFS_I(inode)->rsrc_inode)->rsrc_inode = NULL;
577 iput(HFS_I(inode)->rsrc_inode);
578 }
579}
580
581static int hfs_file_open(struct inode *inode, struct file *file)
582{
583 if (HFS_IS_RSRC(inode))
584 inode = HFS_I(inode)->rsrc_inode;
585 atomic_inc(&HFS_I(inode)->opencnt);
586 return 0;
587}
588
589static int hfs_file_release(struct inode *inode, struct file *file)
590{
591 //struct super_block *sb = inode->i_sb;
592
593 if (HFS_IS_RSRC(inode))
594 inode = HFS_I(inode)->rsrc_inode;
595 if (atomic_dec_and_test(&HFS_I(inode)->opencnt)) {
596 inode_lock(inode);
597 hfs_file_truncate(inode);
598 //if (inode->i_flags & S_DEAD) {
599 // hfs_delete_cat(inode->i_ino, HFSPLUS_SB(sb).hidden_dir, NULL);
600 // hfs_delete_inode(inode);
601 //}
602 inode_unlock(inode);
603 }
604 return 0;
605}
606
607/*
608 * hfs_notify_change()
609 *
610 * Based very closely on fs/msdos/inode.c by Werner Almesberger
611 *
612 * This is the notify_change() field in the super_operations structure
613 * for HFS file systems. The purpose is to take that changes made to
614 * an inode and apply then in a filesystem-dependent manner. In this
615 * case the process has a few of tasks to do:
616 * 1) prevent changes to the i_uid and i_gid fields.
617 * 2) map file permissions to the closest allowable permissions
618 * 3) Since multiple Linux files can share the same on-disk inode under
619 * HFS (for instance the data and resource forks of a file) a change
620 * to permissions must be applied to all other in-core inodes which
621 * correspond to the same HFS file.
622 */
623
624int hfs_inode_setattr(struct mnt_idmap *idmap, struct dentry *dentry,
625 struct iattr *attr)
626{
627 struct inode *inode = d_inode(dentry);
628 struct hfs_sb_info *hsb = HFS_SB(inode->i_sb);
629 int error;
630
631 error = setattr_prepare(&nop_mnt_idmap, dentry,
632 attr); /* basic permission checks */
633 if (error)
634 return error;
635
636 /* no uig/gid changes and limit which mode bits can be set */
637 if (((attr->ia_valid & ATTR_UID) &&
638 (!uid_eq(attr->ia_uid, hsb->s_uid))) ||
639 ((attr->ia_valid & ATTR_GID) &&
640 (!gid_eq(attr->ia_gid, hsb->s_gid))) ||
641 ((attr->ia_valid & ATTR_MODE) &&
642 ((S_ISDIR(inode->i_mode) &&
643 (attr->ia_mode != inode->i_mode)) ||
644 (attr->ia_mode & ~HFS_VALID_MODE_BITS)))) {
645 return hsb->s_quiet ? 0 : error;
646 }
647
648 if (attr->ia_valid & ATTR_MODE) {
649 /* Only the 'w' bits can ever change and only all together. */
650 if (attr->ia_mode & S_IWUSR)
651 attr->ia_mode = inode->i_mode | S_IWUGO;
652 else
653 attr->ia_mode = inode->i_mode & ~S_IWUGO;
654 attr->ia_mode &= S_ISDIR(inode->i_mode) ? ~hsb->s_dir_umask: ~hsb->s_file_umask;
655 }
656
657 if ((attr->ia_valid & ATTR_SIZE) &&
658 attr->ia_size != i_size_read(inode)) {
659 inode_dio_wait(inode);
660
661 error = inode_newsize_ok(inode, attr->ia_size);
662 if (error)
663 return error;
664
665 truncate_setsize(inode, attr->ia_size);
666 hfs_file_truncate(inode);
667 simple_inode_init_ts(inode);
668 }
669
670 setattr_copy(&nop_mnt_idmap, inode, attr);
671 mark_inode_dirty(inode);
672 return 0;
673}
674
675static int hfs_file_fsync(struct file *filp, loff_t start, loff_t end,
676 int datasync)
677{
678 struct inode *inode = filp->f_mapping->host;
679 struct super_block * sb;
680 int ret, err;
681
682 ret = file_write_and_wait_range(filp, start, end);
683 if (ret)
684 return ret;
685 inode_lock(inode);
686
687 /* sync the inode to buffers */
688 ret = write_inode_now(inode, 0);
689
690 /* sync the superblock to buffers */
691 sb = inode->i_sb;
692 flush_delayed_work(&HFS_SB(sb)->mdb_work);
693 /* .. finally sync the buffers to disk */
694 err = sync_blockdev(sb->s_bdev);
695 if (!ret)
696 ret = err;
697 inode_unlock(inode);
698 return ret;
699}
700
701static const struct file_operations hfs_file_operations = {
702 .llseek = generic_file_llseek,
703 .read_iter = generic_file_read_iter,
704 .write_iter = generic_file_write_iter,
705 .mmap_prepare = generic_file_mmap_prepare,
706 .splice_read = filemap_splice_read,
707 .splice_write = iter_file_splice_write,
708 .fsync = hfs_file_fsync,
709 .open = hfs_file_open,
710 .release = hfs_file_release,
711};
712
713static const struct inode_operations hfs_file_inode_operations = {
714 .lookup = hfs_file_lookup,
715 .setattr = hfs_inode_setattr,
716 .listxattr = generic_listxattr,
717};