return 1;
}
-struct buffer_head *
+static struct buffer_head *
__getblk_slow(struct block_device *bdev, sector_t block, int size)
{
/* Size must be multiple of hard sectorsize */
int err;
sector_t block;
sector_t last_block;
- struct buffer_head *bh, *head, *last_bh = NULL;
+ struct buffer_head *bh, *head;
int nr_underway = 0;
BUG_ON(!PageLocked(page));
}
if (test_clear_buffer_dirty(bh)) {
mark_buffer_async_write(bh);
- last_bh = bh;
} else {
unlock_buffer(bh);
}
if (buffer_async_write(bh)) {
submit_bh(WRITE, bh);
nr_underway++;
- if (bh == last_bh)
- break;
}
bh = next;
} while (bh != head);
- bh = head;
unlock_page(page);
err = 0;
if (buffer_mapped(bh) && buffer_dirty(bh)) {
lock_buffer(bh);
mark_buffer_async_write(bh);
- last_bh = bh;
} else {
/*
* The buffer may have been set dirty during
clear_buffer_dirty(bh);
submit_bh(WRITE, bh);
nr_underway++;
- if (bh == last_bh)
- break;
}
bh = next;
} while (bh != head);
- bh = head;
goto done;
}
continue;
if (!buffer_mapped(bh)) {
+ int err = 0;
+
fully_mapped = 0;
if (iblock < lblock) {
- if (get_block(inode, iblock, bh, 0))
+ err = get_block(inode, iblock, bh, 0);
+ if (err)
SetPageError(page);
}
if (!buffer_mapped(bh)) {
memset(kaddr + i * blocksize, 0, blocksize);
flush_dcache_page(page);
kunmap_atomic(kaddr, KM_USER0);
- set_buffer_uptodate(bh);
+ if (!err)
+ set_buffer_uptodate(bh);
continue;
}
/*