5 * Daniel Pirkl <daniel.pirkl@email.cz>
6 * Charles University, Faculty of Mathematics and Physics
9 #include <linux/string.h>
10 #include <linux/slab.h>
11 #include <linux/buffer_head.h>
18 struct ufs_buffer_head * _ubh_bread_ (struct ufs_sb_private_info * uspi,
19 struct super_block *sb, u64 fragment, u64 size)
21 struct ufs_buffer_head * ubh;
24 if (size & ~uspi->s_fmask)
26 count = size >> uspi->s_fshift;
27 if (count > UFS_MAXFRAG)
29 ubh = kmalloc (sizeof (struct ufs_buffer_head), GFP_NOFS);
32 ubh->fragment = fragment;
34 for (i = 0; i < count; i++)
35 if (!(ubh->bh[i] = sb_bread(sb, fragment + i)))
37 for (; i < UFS_MAXFRAG; i++)
41 for (j = 0; j < i; j++)
47 struct ufs_buffer_head * ubh_bread_uspi (struct ufs_sb_private_info * uspi,
48 struct super_block *sb, u64 fragment, u64 size)
52 if (size & ~uspi->s_fmask)
54 count = size >> uspi->s_fshift;
55 if (count <= 0 || count > UFS_MAXFRAG)
57 USPI_UBH(uspi)->fragment = fragment;
58 USPI_UBH(uspi)->count = count;
59 for (i = 0; i < count; i++)
60 if (!(USPI_UBH(uspi)->bh[i] = sb_bread(sb, fragment + i)))
62 for (; i < UFS_MAXFRAG; i++)
63 USPI_UBH(uspi)->bh[i] = NULL;
64 return USPI_UBH(uspi);
66 for (j = 0; j < i; j++)
67 brelse (USPI_UBH(uspi)->bh[j]);
71 void ubh_brelse (struct ufs_buffer_head * ubh)
76 for (i = 0; i < ubh->count; i++)
81 void ubh_brelse_uspi (struct ufs_sb_private_info * uspi)
86 for ( i = 0; i < USPI_UBH(uspi)->count; i++ ) {
87 brelse (USPI_UBH(uspi)->bh[i]);
88 USPI_UBH(uspi)->bh[i] = NULL;
92 void ubh_mark_buffer_dirty (struct ufs_buffer_head * ubh)
97 for ( i = 0; i < ubh->count; i++ )
98 mark_buffer_dirty (ubh->bh[i]);
101 void ubh_mark_buffer_uptodate (struct ufs_buffer_head * ubh, int flag)
107 for ( i = 0; i < ubh->count; i++ )
108 set_buffer_uptodate (ubh->bh[i]);
110 for ( i = 0; i < ubh->count; i++ )
111 clear_buffer_uptodate (ubh->bh[i]);
115 void ubh_sync_block(struct ufs_buffer_head *ubh)
120 for (i = 0; i < ubh->count; i++)
121 write_dirty_buffer(ubh->bh[i], WRITE);
123 for (i = 0; i < ubh->count; i++)
124 wait_on_buffer(ubh->bh[i]);
128 void ubh_bforget (struct ufs_buffer_head * ubh)
133 for ( i = 0; i < ubh->count; i++ ) if ( ubh->bh[i] )
134 bforget (ubh->bh[i]);
137 int ubh_buffer_dirty (struct ufs_buffer_head * ubh)
143 for ( i = 0; i < ubh->count; i++ )
144 result |= buffer_dirty(ubh->bh[i]);
148 void _ubh_ubhcpymem_(struct ufs_sb_private_info * uspi,
149 unsigned char * mem, struct ufs_buffer_head * ubh, unsigned size)
152 if (size > (ubh->count << uspi->s_fshift))
153 size = ubh->count << uspi->s_fshift;
156 len = min_t(unsigned int, size, uspi->s_fsize);
157 memcpy (mem, ubh->bh[bhno]->b_data, len);
158 mem += uspi->s_fsize;
164 void _ubh_memcpyubh_(struct ufs_sb_private_info * uspi,
165 struct ufs_buffer_head * ubh, unsigned char * mem, unsigned size)
168 if (size > (ubh->count << uspi->s_fshift))
169 size = ubh->count << uspi->s_fshift;
172 len = min_t(unsigned int, size, uspi->s_fsize);
173 memcpy (ubh->bh[bhno]->b_data, mem, len);
174 mem += uspi->s_fsize;
181 ufs_get_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi)
186 if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
187 fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[1]);
189 fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[0]);
190 switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
193 if ((fs32 & 0xffff0000) == 0 ||
194 (fs32 & 0xffff0000) == 0xffff0000)
195 dev = old_decode_dev(fs32 & 0x7fff);
197 dev = MKDEV(sysv_major(fs32), sysv_minor(fs32));
201 dev = old_decode_dev(fs32);
208 ufs_set_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi, dev_t dev)
212 switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
215 fs32 = sysv_encode_dev(dev);
216 if ((fs32 & 0xffff8000) == 0) {
217 fs32 = old_encode_dev(dev);
222 fs32 = old_encode_dev(dev);
225 if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
226 ufsi->i_u1.i_data[1] = cpu_to_fs32(sb, fs32);
228 ufsi->i_u1.i_data[0] = cpu_to_fs32(sb, fs32);
232 * ufs_get_locked_page() - locate, pin and lock a pagecache page, if not exist
234 * @mapping: the address_space to search
235 * @index: the page index
237 * Locates the desired pagecache page, if not exist we'll read it,
238 * locks it, increments its reference
239 * count and returns its address.
243 struct page *ufs_get_locked_page(struct address_space *mapping,
248 page = find_lock_page(mapping, index);
250 page = read_mapping_page(mapping, index, NULL);
253 printk(KERN_ERR "ufs_change_blocknr: "
254 "read_mapping_page error: ino %lu, index: %lu\n",
255 mapping->host->i_ino, index);
261 if (unlikely(page->mapping == NULL)) {
262 /* Truncate got there first */
264 page_cache_release(page);
269 if (!PageUptodate(page) || PageError(page)) {
271 page_cache_release(page);
273 printk(KERN_ERR "ufs_change_blocknr: "
274 "can not read page: ino %lu, index: %lu\n",
275 mapping->host->i_ino, index);
277 page = ERR_PTR(-EIO);