If there are multi segments in one section, we will read those SSA blocks which
have contiguous address one by one in f2fs_gc. It may lost performance, let's
read ahead SSA blocks by merge multi read request.
Signed-off-by: Chao Yu <chao2.yu@samsung.com>
Signed-off-by: Jaegeuk Kim <jaegeuk.kim@samsung.com>
return NM_I(sbi)->max_nid / NAT_ENTRY_PER_BLOCK;
case META_SIT:
return SIT_BLK_CNT(sbi);
return NM_I(sbi)->max_nid / NAT_ENTRY_PER_BLOCK;
case META_SIT:
return SIT_BLK_CNT(sbi);
case META_CP:
return 0;
default:
case META_CP:
return 0;
default:
- * Readahead CP/NAT/SIT pages
+ * Readahead CP/NAT/SIT/SSA pages
*/
int ra_meta_pages(struct f2fs_sb_info *sbi, int start, int nrpages, int type)
{
*/
int ra_meta_pages(struct f2fs_sb_info *sbi, int start, int nrpages, int type)
{
goto out;
prev_blk_addr = blk_addr;
break;
goto out;
prev_blk_addr = blk_addr;
break;
- /* get cp block addr */
+ /* get ssa/cp block addr */
blk_addr = blkno;
break;
default:
blk_addr = blkno;
break;
default:
- * For CP/NAT/SIT readahead
+ * For CP/NAT/SIT/SSA readahead
*/
enum {
META_CP,
META_NAT,
*/
enum {
META_CP,
META_NAT,
};
/* for the list of orphan inodes */
};
/* for the list of orphan inodes */
+ /* readahead multi ssa blocks those have contiguous address */
+ if (sbi->segs_per_sec > 1)
+ ra_meta_pages(sbi, GET_SUM_BLOCK(sbi, segno), sbi->segs_per_sec,
+ META_SSA);
+
for (i = 0; i < sbi->segs_per_sec; i++)
do_garbage_collect(sbi, segno + i, &ilist, gc_type);
for (i = 0; i < sbi->segs_per_sec; i++)
do_garbage_collect(sbi, segno + i, &ilist, gc_type);