4 * Copyright (c) 2013 Samsung Electronics Co., Ltd.
5 * http://www.samsung.com/
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
14 #ifdef HAVE_LINUX_POSIX_ACL_H
15 #include <linux/posix_acl.h>
21 #ifndef ACL_UNDEFINED_TAG
22 #define ACL_UNDEFINED_TAG (0x00)
23 #define ACL_USER_OBJ (0x01)
24 #define ACL_USER (0x02)
25 #define ACL_GROUP_OBJ (0x04)
26 #define ACL_GROUP (0x08)
27 #define ACL_MASK (0x10)
28 #define ACL_OTHER (0x20)
31 u32 get_free_segments(struct f2fs_sb_info *sbi)
35 for (i = 0; i < TOTAL_SEGS(sbi); i++) {
36 struct seg_entry *se = get_seg_entry(sbi, i);
38 if (se->valid_blocks == 0x0 && !IS_CUR_SEGNO(sbi, i))
44 void update_free_segments(struct f2fs_sb_info *sbi)
46 char *progress = "-*|*-";
52 MSG(0, "\r [ %c ] Free segments: 0x%x", progress[i % 5], get_free_segments(sbi));
57 #if defined(HAVE_LINUX_POSIX_ACL_H) || defined(HAVE_SYS_ACL_H)
58 static void print_acl(const u8 *value, int size)
60 const struct f2fs_acl_header *hdr = (struct f2fs_acl_header *)value;
61 const struct f2fs_acl_entry *entry = (struct f2fs_acl_entry *)(hdr + 1);
62 const u8 *end = value + size;
65 if (hdr->a_version != cpu_to_le32(F2FS_ACL_VERSION)) {
66 MSG(0, "Invalid ACL version [0x%x : 0x%x]\n",
67 le32_to_cpu(hdr->a_version), F2FS_ACL_VERSION);
71 count = f2fs_acl_count(size);
73 MSG(0, "Invalid ACL value size %d\n", size);
77 for (i = 0; i < count; i++) {
78 if ((u8 *)entry > end) {
79 MSG(0, "Invalid ACL entries count %d\n", count);
83 switch (le16_to_cpu(entry->e_tag)) {
88 MSG(0, "tag:0x%x perm:0x%x\n",
89 le16_to_cpu(entry->e_tag),
90 le16_to_cpu(entry->e_perm));
91 entry = (struct f2fs_acl_entry *)((char *)entry +
92 sizeof(struct f2fs_acl_entry_short));
95 MSG(0, "tag:0x%x perm:0x%x uid:%u\n",
96 le16_to_cpu(entry->e_tag),
97 le16_to_cpu(entry->e_perm),
98 le32_to_cpu(entry->e_id));
99 entry = (struct f2fs_acl_entry *)((char *)entry +
100 sizeof(struct f2fs_acl_entry));
103 MSG(0, "tag:0x%x perm:0x%x gid:%u\n",
104 le16_to_cpu(entry->e_tag),
105 le16_to_cpu(entry->e_perm),
106 le32_to_cpu(entry->e_id));
107 entry = (struct f2fs_acl_entry *)((char *)entry +
108 sizeof(struct f2fs_acl_entry));
111 MSG(0, "Unknown ACL tag 0x%x\n",
112 le16_to_cpu(entry->e_tag));
117 #endif /* HAVE_LINUX_POSIX_ACL_H || HAVE_SYS_ACL_H */
119 static void print_xattr_entry(const struct f2fs_xattr_entry *ent)
121 const u8 *value = (const u8 *)&ent->e_name[ent->e_name_len];
122 const int size = le16_to_cpu(ent->e_value_size);
123 const struct fscrypt_context *ctx;
126 MSG(0, "\nxattr: e_name_index:%d e_name:", ent->e_name_index);
127 for (i = 0; i < ent->e_name_len; i++)
128 MSG(0, "%c", ent->e_name[i]);
129 MSG(0, " e_name_len:%d e_value_size:%d e_value:\n",
130 ent->e_name_len, size);
132 switch (ent->e_name_index) {
133 #if defined(HAVE_LINUX_POSIX_ACL_H) || defined(HAVE_SYS_ACL_H)
134 case F2FS_XATTR_INDEX_POSIX_ACL_ACCESS:
135 case F2FS_XATTR_INDEX_POSIX_ACL_DEFAULT:
136 print_acl(value, size);
139 case F2FS_XATTR_INDEX_ENCRYPTION:
140 ctx = (const struct fscrypt_context *)value;
141 if (size != sizeof(*ctx) ||
142 ctx->format != FS_ENCRYPTION_CONTEXT_FORMAT_V1)
144 MSG(0, "format: %d\n", ctx->format);
145 MSG(0, "contents_encryption_mode: 0x%x\n", ctx->contents_encryption_mode);
146 MSG(0, "filenames_encryption_mode: 0x%x\n", ctx->filenames_encryption_mode);
147 MSG(0, "flags: 0x%x\n", ctx->flags);
148 MSG(0, "master_key_descriptor: ");
149 for (i = 0; i < FS_KEY_DESCRIPTOR_SIZE; i++)
150 MSG(0, "%02X", ctx->master_key_descriptor[i]);
152 for (i = 0; i < FS_KEY_DERIVATION_NONCE_SIZE; i++)
153 MSG(0, "%02X", ctx->nonce[i]);
157 for (i = 0; i < size; i++)
158 MSG(0, "%02X", value[i]);
162 void print_inode_info(struct f2fs_sb_info *sbi,
163 struct f2fs_node *node, int name)
165 struct f2fs_inode *inode = &node->i;
167 struct f2fs_xattr_entry *ent;
168 char en[F2FS_PRINT_NAMELEN];
170 u32 namelen = le32_to_cpu(inode->i_namelen);
171 int enc_name = file_enc_name(inode);
172 int ofs = __get_extra_isize(inode);
174 pretty_print_filename(inode->i_name, namelen, en, enc_name);
176 MSG(0, " - File name : %s%s\n", en,
177 enc_name ? " <encrypted>" : "");
178 setlocale(LC_ALL, "");
179 MSG(0, " - File size : %'llu (bytes)\n",
180 le64_to_cpu(inode->i_size));
184 DISP_u32(inode, i_mode);
185 DISP_u32(inode, i_advise);
186 DISP_u32(inode, i_uid);
187 DISP_u32(inode, i_gid);
188 DISP_u32(inode, i_links);
189 DISP_u64(inode, i_size);
190 DISP_u64(inode, i_blocks);
192 DISP_u64(inode, i_atime);
193 DISP_u32(inode, i_atime_nsec);
194 DISP_u64(inode, i_ctime);
195 DISP_u32(inode, i_ctime_nsec);
196 DISP_u64(inode, i_mtime);
197 DISP_u32(inode, i_mtime_nsec);
199 DISP_u32(inode, i_generation);
200 DISP_u32(inode, i_current_depth);
201 DISP_u32(inode, i_xattr_nid);
202 DISP_u32(inode, i_flags);
203 DISP_u32(inode, i_inline);
204 DISP_u32(inode, i_pino);
205 DISP_u32(inode, i_dir_level);
208 DISP_u32(inode, i_namelen);
209 printf("%-30s\t\t[%s]\n", "i_name", en);
212 printf("i_ext: fofs:%x blkaddr:%x len:%x\n",
213 le32_to_cpu(inode->i_ext.fofs),
214 le32_to_cpu(inode->i_ext.blk_addr),
215 le32_to_cpu(inode->i_ext.len));
217 if (c.feature & cpu_to_le32(F2FS_FEATURE_EXTRA_ATTR)) {
218 DISP_u16(inode, i_extra_isize);
219 if (c.feature & cpu_to_le32(F2FS_FEATURE_FLEXIBLE_INLINE_XATTR))
220 DISP_u16(inode, i_inline_xattr_size);
221 if (c.feature & cpu_to_le32(F2FS_FEATURE_PRJQUOTA))
222 DISP_u32(inode, i_projid);
223 if (c.feature & cpu_to_le32(F2FS_FEATURE_INODE_CHKSUM))
224 DISP_u32(inode, i_inode_checksum);
225 if (c.feature & cpu_to_le32(F2FS_FEATURE_INODE_CRTIME)) {
226 DISP_u64(inode, i_crtime);
227 DISP_u32(inode, i_crtime_nsec);
231 DISP_u32(inode, i_addr[ofs]); /* Pointers to data blocks */
232 DISP_u32(inode, i_addr[ofs + 1]); /* Pointers to data blocks */
233 DISP_u32(inode, i_addr[ofs + 2]); /* Pointers to data blocks */
234 DISP_u32(inode, i_addr[ofs + 3]); /* Pointers to data blocks */
236 for (i = ofs + 3; i < ADDRS_PER_INODE(inode); i++) {
237 if (inode->i_addr[i] == 0x0)
239 printf("i_addr[0x%x] points data block\t\t[0x%4x]\n",
240 i, le32_to_cpu(inode->i_addr[i]));
243 DISP_u32(inode, i_nid[0]); /* direct */
244 DISP_u32(inode, i_nid[1]); /* direct */
245 DISP_u32(inode, i_nid[2]); /* indirect */
246 DISP_u32(inode, i_nid[3]); /* indirect */
247 DISP_u32(inode, i_nid[4]); /* double indirect */
249 xattr_addr = read_all_xattrs(sbi, node);
250 list_for_each_xattr(ent, xattr_addr) {
251 print_xattr_entry(ent);
258 void print_node_info(struct f2fs_sb_info *sbi,
259 struct f2fs_node *node_block, int verbose)
261 nid_t ino = le32_to_cpu(node_block->footer.ino);
262 nid_t nid = le32_to_cpu(node_block->footer.nid);
265 DBG(verbose, "Node ID [0x%x:%u] is inode\n", nid, nid);
266 print_inode_info(sbi, node_block, verbose);
269 u32 *dump_blk = (u32 *)node_block;
271 "Node ID [0x%x:%u] is direct node or indirect node.\n",
273 for (i = 0; i <= 10; i++)
274 MSG(verbose, "[%d]\t\t\t[0x%8x : %d]\n",
275 i, dump_blk[i], dump_blk[i]);
279 static void DISP_label(u_int16_t *name)
281 char buffer[MAX_VOLUME_NAME];
283 utf16_to_utf8(buffer, name, MAX_VOLUME_NAME, MAX_VOLUME_NAME);
284 printf("%-30s" "\t\t[%s]\n", "volum_name", buffer);
287 void print_raw_sb_info(struct f2fs_super_block *sb)
293 printf("+--------------------------------------------------------+\n");
294 printf("| Super block |\n");
295 printf("+--------------------------------------------------------+\n");
298 DISP_u32(sb, major_ver);
300 DISP_label(sb->volume_name);
302 DISP_u32(sb, minor_ver);
303 DISP_u32(sb, log_sectorsize);
304 DISP_u32(sb, log_sectors_per_block);
306 DISP_u32(sb, log_blocksize);
307 DISP_u32(sb, log_blocks_per_seg);
308 DISP_u32(sb, segs_per_sec);
309 DISP_u32(sb, secs_per_zone);
310 DISP_u32(sb, checksum_offset);
311 DISP_u64(sb, block_count);
313 DISP_u32(sb, section_count);
314 DISP_u32(sb, segment_count);
315 DISP_u32(sb, segment_count_ckpt);
316 DISP_u32(sb, segment_count_sit);
317 DISP_u32(sb, segment_count_nat);
319 DISP_u32(sb, segment_count_ssa);
320 DISP_u32(sb, segment_count_main);
321 DISP_u32(sb, segment0_blkaddr);
323 DISP_u32(sb, cp_blkaddr);
324 DISP_u32(sb, sit_blkaddr);
325 DISP_u32(sb, nat_blkaddr);
326 DISP_u32(sb, ssa_blkaddr);
327 DISP_u32(sb, main_blkaddr);
329 DISP_u32(sb, root_ino);
330 DISP_u32(sb, node_ino);
331 DISP_u32(sb, meta_ino);
332 DISP_u32(sb, cp_payload);
334 DISP("%-.256s", sb, version);
338 void print_ckpt_info(struct f2fs_sb_info *sbi)
340 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
346 printf("+--------------------------------------------------------+\n");
347 printf("| Checkpoint |\n");
348 printf("+--------------------------------------------------------+\n");
350 DISP_u64(cp, checkpoint_ver);
351 DISP_u64(cp, user_block_count);
352 DISP_u64(cp, valid_block_count);
353 DISP_u32(cp, rsvd_segment_count);
354 DISP_u32(cp, overprov_segment_count);
355 DISP_u32(cp, free_segment_count);
357 DISP_u32(cp, alloc_type[CURSEG_HOT_NODE]);
358 DISP_u32(cp, alloc_type[CURSEG_WARM_NODE]);
359 DISP_u32(cp, alloc_type[CURSEG_COLD_NODE]);
360 DISP_u32(cp, cur_node_segno[0]);
361 DISP_u32(cp, cur_node_segno[1]);
362 DISP_u32(cp, cur_node_segno[2]);
364 DISP_u32(cp, cur_node_blkoff[0]);
365 DISP_u32(cp, cur_node_blkoff[1]);
366 DISP_u32(cp, cur_node_blkoff[2]);
369 DISP_u32(cp, alloc_type[CURSEG_HOT_DATA]);
370 DISP_u32(cp, alloc_type[CURSEG_WARM_DATA]);
371 DISP_u32(cp, alloc_type[CURSEG_COLD_DATA]);
372 DISP_u32(cp, cur_data_segno[0]);
373 DISP_u32(cp, cur_data_segno[1]);
374 DISP_u32(cp, cur_data_segno[2]);
376 DISP_u32(cp, cur_data_blkoff[0]);
377 DISP_u32(cp, cur_data_blkoff[1]);
378 DISP_u32(cp, cur_data_blkoff[2]);
380 DISP_u32(cp, ckpt_flags);
381 DISP_u32(cp, cp_pack_total_block_count);
382 DISP_u32(cp, cp_pack_start_sum);
383 DISP_u32(cp, valid_node_count);
384 DISP_u32(cp, valid_inode_count);
385 DISP_u32(cp, next_free_nid);
386 DISP_u32(cp, sit_ver_bitmap_bytesize);
387 DISP_u32(cp, nat_ver_bitmap_bytesize);
388 DISP_u32(cp, checksum_offset);
389 DISP_u64(cp, elapsed_time);
391 DISP_u32(cp, sit_nat_version_bitmap[0]);
395 void print_cp_state(u32 flag)
397 MSG(0, "Info: checkpoint state = %x : ", flag);
398 if (flag & CP_QUOTA_NEED_FSCK_FLAG)
399 MSG(0, "%s", " quota_need_fsck");
400 if (flag & CP_LARGE_NAT_BITMAP_FLAG)
401 MSG(0, "%s", " large_nat_bitmap");
402 if (flag & CP_NOCRC_RECOVERY_FLAG)
403 MSG(0, "%s", " allow_nocrc");
404 if (flag & CP_TRIMMED_FLAG)
405 MSG(0, "%s", " trimmed");
406 if (flag & CP_NAT_BITS_FLAG)
407 MSG(0, "%s", " nat_bits");
408 if (flag & CP_CRC_RECOVERY_FLAG)
409 MSG(0, "%s", " crc");
410 if (flag & CP_FASTBOOT_FLAG)
411 MSG(0, "%s", " fastboot");
412 if (flag & CP_FSCK_FLAG)
413 MSG(0, "%s", " fsck");
414 if (flag & CP_ERROR_FLAG)
415 MSG(0, "%s", " error");
416 if (flag & CP_COMPACT_SUM_FLAG)
417 MSG(0, "%s", " compacted_summary");
418 if (flag & CP_ORPHAN_PRESENT_FLAG)
419 MSG(0, "%s", " orphan_inodes");
420 if (flag & CP_DISABLED_FLAG)
421 MSG(0, "%s", " disabled");
422 if (flag & CP_UMOUNT_FLAG)
423 MSG(0, "%s", " unmount");
425 MSG(0, "%s", " sudden-power-off");
429 void print_sb_state(struct f2fs_super_block *sb)
431 __le32 f = sb->feature;
434 MSG(0, "Info: superblock features = %x : ", f);
435 if (f & cpu_to_le32(F2FS_FEATURE_ENCRYPT)) {
436 MSG(0, "%s", " encrypt");
438 if (f & cpu_to_le32(F2FS_FEATURE_VERITY)) {
439 MSG(0, "%s", " verity");
441 if (f & cpu_to_le32(F2FS_FEATURE_BLKZONED)) {
442 MSG(0, "%s", " blkzoned");
444 if (f & cpu_to_le32(F2FS_FEATURE_EXTRA_ATTR)) {
445 MSG(0, "%s", " extra_attr");
447 if (f & cpu_to_le32(F2FS_FEATURE_PRJQUOTA)) {
448 MSG(0, "%s", " project_quota");
450 if (f & cpu_to_le32(F2FS_FEATURE_INODE_CHKSUM)) {
451 MSG(0, "%s", " inode_checksum");
453 if (f & cpu_to_le32(F2FS_FEATURE_FLEXIBLE_INLINE_XATTR)) {
454 MSG(0, "%s", " flexible_inline_xattr");
456 if (f & cpu_to_le32(F2FS_FEATURE_QUOTA_INO)) {
457 MSG(0, "%s", " quota_ino");
459 if (f & cpu_to_le32(F2FS_FEATURE_INODE_CRTIME)) {
460 MSG(0, "%s", " inode_crtime");
462 if (f & cpu_to_le32(F2FS_FEATURE_LOST_FOUND)) {
463 MSG(0, "%s", " lost_found");
465 if (f & cpu_to_le32(F2FS_FEATURE_SB_CHKSUM)) {
466 MSG(0, "%s", " sb_checksum");
469 MSG(0, "Info: superblock encrypt level = %d, salt = ",
470 sb->encryption_level);
471 for (i = 0; i < 16; i++)
472 MSG(0, "%02x", sb->encrypt_pw_salt[i]);
476 void update_superblock(struct f2fs_super_block *sb, int sb_mask)
480 u32 old_crc, new_crc;
482 buf = calloc(BLOCK_SZ, 1);
485 if (get_sb(feature) & F2FS_FEATURE_SB_CHKSUM) {
486 old_crc = get_sb(crc);
487 new_crc = f2fs_cal_crc32(F2FS_SUPER_MAGIC, sb,
489 set_sb(crc, new_crc);
490 MSG(1, "Info: SB CRC is updated (0x%x -> 0x%x)\n",
494 memcpy(buf + F2FS_SUPER_OFFSET, sb, sizeof(*sb));
495 for (addr = SB0_ADDR; addr < SB_MAX_ADDR; addr++) {
496 if (SB_MASK(addr) & sb_mask) {
497 ret = dev_write_block(buf, addr);
503 DBG(0, "Info: Done to update superblock\n");
506 static inline int sanity_check_area_boundary(struct f2fs_super_block *sb,
507 enum SB_ADDR sb_addr)
509 u32 segment0_blkaddr = get_sb(segment0_blkaddr);
510 u32 cp_blkaddr = get_sb(cp_blkaddr);
511 u32 sit_blkaddr = get_sb(sit_blkaddr);
512 u32 nat_blkaddr = get_sb(nat_blkaddr);
513 u32 ssa_blkaddr = get_sb(ssa_blkaddr);
514 u32 main_blkaddr = get_sb(main_blkaddr);
515 u32 segment_count_ckpt = get_sb(segment_count_ckpt);
516 u32 segment_count_sit = get_sb(segment_count_sit);
517 u32 segment_count_nat = get_sb(segment_count_nat);
518 u32 segment_count_ssa = get_sb(segment_count_ssa);
519 u32 segment_count_main = get_sb(segment_count_main);
520 u32 segment_count = get_sb(segment_count);
521 u32 log_blocks_per_seg = get_sb(log_blocks_per_seg);
522 u64 main_end_blkaddr = main_blkaddr +
523 (segment_count_main << log_blocks_per_seg);
524 u64 seg_end_blkaddr = segment0_blkaddr +
525 (segment_count << log_blocks_per_seg);
527 if (segment0_blkaddr != cp_blkaddr) {
528 MSG(0, "\tMismatch segment0(%u) cp_blkaddr(%u)\n",
529 segment0_blkaddr, cp_blkaddr);
533 if (cp_blkaddr + (segment_count_ckpt << log_blocks_per_seg) !=
535 MSG(0, "\tWrong CP boundary, start(%u) end(%u) blocks(%u)\n",
536 cp_blkaddr, sit_blkaddr,
537 segment_count_ckpt << log_blocks_per_seg);
541 if (sit_blkaddr + (segment_count_sit << log_blocks_per_seg) !=
543 MSG(0, "\tWrong SIT boundary, start(%u) end(%u) blocks(%u)\n",
544 sit_blkaddr, nat_blkaddr,
545 segment_count_sit << log_blocks_per_seg);
549 if (nat_blkaddr + (segment_count_nat << log_blocks_per_seg) !=
551 MSG(0, "\tWrong NAT boundary, start(%u) end(%u) blocks(%u)\n",
552 nat_blkaddr, ssa_blkaddr,
553 segment_count_nat << log_blocks_per_seg);
557 if (ssa_blkaddr + (segment_count_ssa << log_blocks_per_seg) !=
559 MSG(0, "\tWrong SSA boundary, start(%u) end(%u) blocks(%u)\n",
560 ssa_blkaddr, main_blkaddr,
561 segment_count_ssa << log_blocks_per_seg);
565 if (main_end_blkaddr > seg_end_blkaddr) {
566 MSG(0, "\tWrong MAIN_AREA, start(%u) end(%u) block(%u)\n",
569 (segment_count << log_blocks_per_seg),
570 segment_count_main << log_blocks_per_seg);
572 } else if (main_end_blkaddr < seg_end_blkaddr) {
573 set_sb(segment_count, (main_end_blkaddr -
574 segment0_blkaddr) >> log_blocks_per_seg);
576 update_superblock(sb, SB_MASK(sb_addr));
577 MSG(0, "Info: Fix alignment: start(%u) end(%u) block(%u)\n",
580 (segment_count << log_blocks_per_seg),
581 segment_count_main << log_blocks_per_seg);
586 static int verify_sb_chksum(struct f2fs_super_block *sb)
588 if (SB_CHKSUM_OFFSET != get_sb(checksum_offset)) {
589 MSG(0, "\tInvalid SB CRC offset: %u\n",
590 get_sb(checksum_offset));
593 if (f2fs_crc_valid(get_sb(crc), sb,
594 get_sb(checksum_offset))) {
595 MSG(0, "\tInvalid SB CRC: 0x%x\n", get_sb(crc));
601 int sanity_check_raw_super(struct f2fs_super_block *sb, enum SB_ADDR sb_addr)
603 unsigned int blocksize;
605 if ((get_sb(feature) & F2FS_FEATURE_SB_CHKSUM) &&
606 verify_sb_chksum(sb))
609 if (F2FS_SUPER_MAGIC != get_sb(magic))
612 if (F2FS_BLKSIZE != PAGE_CACHE_SIZE)
615 blocksize = 1 << get_sb(log_blocksize);
616 if (F2FS_BLKSIZE != blocksize)
619 /* check log blocks per segment */
620 if (get_sb(log_blocks_per_seg) != 9)
623 /* Currently, support 512/1024/2048/4096 bytes sector size */
624 if (get_sb(log_sectorsize) > F2FS_MAX_LOG_SECTOR_SIZE ||
625 get_sb(log_sectorsize) < F2FS_MIN_LOG_SECTOR_SIZE)
628 if (get_sb(log_sectors_per_block) + get_sb(log_sectorsize) !=
629 F2FS_MAX_LOG_SECTOR_SIZE)
632 /* check reserved ino info */
633 if (get_sb(node_ino) != 1 || get_sb(meta_ino) != 2 ||
634 get_sb(root_ino) != 3)
637 /* Check zoned block device feature */
638 if (c.devices[0].zoned_model == F2FS_ZONED_HM &&
639 !(sb->feature & cpu_to_le32(F2FS_FEATURE_BLKZONED))) {
640 MSG(0, "\tMissing zoned block device feature\n");
644 if (get_sb(segment_count) > F2FS_MAX_SEGMENT)
647 if (sanity_check_area_boundary(sb, sb_addr))
652 int validate_super_block(struct f2fs_sb_info *sbi, enum SB_ADDR sb_addr)
654 char buf[F2FS_BLKSIZE];
656 sbi->raw_super = malloc(sizeof(struct f2fs_super_block));
660 if (dev_read_block(buf, sb_addr))
663 memcpy(sbi->raw_super, buf + F2FS_SUPER_OFFSET,
664 sizeof(struct f2fs_super_block));
666 if (!sanity_check_raw_super(sbi->raw_super, sb_addr)) {
667 /* get kernel version */
669 dev_read_version(c.version, 0, VERSION_LEN);
670 get_kernel_version(c.version);
672 get_kernel_uname_version(c.version);
675 /* build sb version */
676 memcpy(c.sb_version, sbi->raw_super->version, VERSION_LEN);
677 get_kernel_version(c.sb_version);
678 memcpy(c.init_version, sbi->raw_super->init_version, VERSION_LEN);
679 get_kernel_version(c.init_version);
681 MSG(0, "Info: MKFS version\n \"%s\"\n", c.init_version);
682 MSG(0, "Info: FSCK version\n from \"%s\"\n to \"%s\"\n",
683 c.sb_version, c.version);
684 if (memcmp(c.sb_version, c.version, VERSION_LEN)) {
685 memcpy(sbi->raw_super->version,
686 c.version, VERSION_LEN);
687 update_superblock(sbi->raw_super, SB_MASK(sb_addr));
692 print_sb_state(sbi->raw_super);
696 free(sbi->raw_super);
697 sbi->raw_super = NULL;
698 MSG(0, "\tCan't find a valid F2FS superblock at 0x%x\n", sb_addr);
703 int init_sb_info(struct f2fs_sb_info *sbi)
705 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
709 sbi->log_sectors_per_block = get_sb(log_sectors_per_block);
710 sbi->log_blocksize = get_sb(log_blocksize);
711 sbi->blocksize = 1 << sbi->log_blocksize;
712 sbi->log_blocks_per_seg = get_sb(log_blocks_per_seg);
713 sbi->blocks_per_seg = 1 << sbi->log_blocks_per_seg;
714 sbi->segs_per_sec = get_sb(segs_per_sec);
715 sbi->secs_per_zone = get_sb(secs_per_zone);
716 sbi->total_sections = get_sb(section_count);
717 sbi->total_node_count = (get_sb(segment_count_nat) / 2) *
718 sbi->blocks_per_seg * NAT_ENTRY_PER_BLOCK;
719 sbi->root_ino_num = get_sb(root_ino);
720 sbi->node_ino_num = get_sb(node_ino);
721 sbi->meta_ino_num = get_sb(meta_ino);
722 sbi->cur_victim_sec = NULL_SEGNO;
724 for (i = 0; i < MAX_DEVICES; i++) {
725 if (!sb->devs[i].path[0])
729 c.devices[i].path = strdup((char *)sb->devs[i].path);
730 if (get_device_info(i))
733 ASSERT(!strcmp((char *)sb->devs[i].path,
734 (char *)c.devices[i].path));
737 c.devices[i].total_segments =
738 le32_to_cpu(sb->devs[i].total_segments);
740 c.devices[i].start_blkaddr =
741 c.devices[i - 1].end_blkaddr + 1;
742 c.devices[i].end_blkaddr = c.devices[i].start_blkaddr +
743 c.devices[i].total_segments *
746 c.devices[i].end_blkaddr += get_sb(segment0_blkaddr);
749 MSG(0, "Info: Device[%d] : %s blkaddr = %"PRIx64"--%"PRIx64"\n",
750 i, c.devices[i].path,
751 c.devices[i].start_blkaddr,
752 c.devices[i].end_blkaddr);
755 total_sectors = get_sb(block_count) << sbi->log_sectors_per_block;
756 MSG(0, "Info: total FS sectors = %"PRIu64" (%"PRIu64" MB)\n",
757 total_sectors, total_sectors >>
758 (20 - get_sb(log_sectorsize)));
762 static int verify_checksum_chksum(struct f2fs_checkpoint *cp)
764 unsigned int chksum_offset = get_cp(checksum_offset);
765 unsigned int crc, cal_crc;
767 if (chksum_offset < CP_MIN_CHKSUM_OFFSET ||
768 chksum_offset > CP_CHKSUM_OFFSET) {
769 MSG(0, "\tInvalid CP CRC offset: %u\n", chksum_offset);
773 crc = le32_to_cpu(*(__le32 *)((unsigned char *)cp + chksum_offset));
774 cal_crc = f2fs_checkpoint_chksum(cp);
775 if (cal_crc != crc) {
776 MSG(0, "\tInvalid CP CRC: offset:%u, crc:0x%x, calc:0x%x\n",
777 chksum_offset, crc, cal_crc);
783 static void *get_checkpoint_version(block_t cp_addr)
787 cp_page = malloc(PAGE_SIZE);
790 if (dev_read_block(cp_page, cp_addr) < 0)
793 if (verify_checksum_chksum((struct f2fs_checkpoint *)cp_page))
801 void *validate_checkpoint(struct f2fs_sb_info *sbi, block_t cp_addr,
802 unsigned long long *version)
804 void *cp_page_1, *cp_page_2;
805 struct f2fs_checkpoint *cp;
806 unsigned long long cur_version = 0, pre_version = 0;
808 /* Read the 1st cp block in this CP pack */
809 cp_page_1 = get_checkpoint_version(cp_addr);
813 cp = (struct f2fs_checkpoint *)cp_page_1;
814 if (get_cp(cp_pack_total_block_count) > sbi->blocks_per_seg)
817 pre_version = get_cp(checkpoint_ver);
819 /* Read the 2nd cp block in this CP pack */
820 cp_addr += get_cp(cp_pack_total_block_count) - 1;
821 cp_page_2 = get_checkpoint_version(cp_addr);
825 cp = (struct f2fs_checkpoint *)cp_page_2;
826 cur_version = get_cp(checkpoint_ver);
828 if (cur_version == pre_version) {
829 *version = cur_version;
840 int get_valid_checkpoint(struct f2fs_sb_info *sbi)
842 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
843 void *cp1, *cp2, *cur_page;
844 unsigned long blk_size = sbi->blocksize;
845 unsigned long long cp1_version = 0, cp2_version = 0, version;
846 unsigned long long cp_start_blk_no;
847 unsigned int cp_payload, cp_blks;
850 cp_payload = get_sb(cp_payload);
851 if (cp_payload > F2FS_BLK_ALIGN(MAX_SIT_BITMAP_SIZE))
854 cp_blks = 1 + cp_payload;
855 sbi->ckpt = malloc(cp_blks * blk_size);
859 * Finding out valid cp block involves read both
860 * sets( cp pack1 and cp pack 2)
862 cp_start_blk_no = get_sb(cp_blkaddr);
863 cp1 = validate_checkpoint(sbi, cp_start_blk_no, &cp1_version);
865 /* The second checkpoint pack should start at the next segment */
866 cp_start_blk_no += 1 << get_sb(log_blocks_per_seg);
867 cp2 = validate_checkpoint(sbi, cp_start_blk_no, &cp2_version);
870 if (ver_after(cp2_version, cp1_version)) {
873 version = cp2_version;
877 version = cp1_version;
882 version = cp1_version;
886 version = cp2_version;
890 MSG(0, "Info: CKPT version = %llx\n", version);
892 memcpy(sbi->ckpt, cur_page, blk_size);
896 unsigned long long cp_blk_no;
898 cp_blk_no = get_sb(cp_blkaddr);
900 cp_blk_no += 1 << get_sb(log_blocks_per_seg);
902 /* copy sit bitmap */
903 for (i = 1; i < cp_blks; i++) {
904 unsigned char *ckpt = (unsigned char *)sbi->ckpt;
905 ret = dev_read_block(cur_page, cp_blk_no + i);
907 memcpy(ckpt + i * blk_size, cur_page, blk_size);
922 int sanity_check_ckpt(struct f2fs_sb_info *sbi)
924 unsigned int total, fsmeta;
925 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
926 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
927 unsigned int ovp_segments, reserved_segments;
928 unsigned int main_segs, blocks_per_seg;
929 unsigned int sit_segs, nat_segs;
930 unsigned int sit_bitmap_size, nat_bitmap_size;
931 unsigned int log_blocks_per_seg;
932 unsigned int segment_count_main;
933 unsigned int cp_pack_start_sum, cp_payload;
934 block_t user_block_count;
937 total = get_sb(segment_count);
938 fsmeta = get_sb(segment_count_ckpt);
939 sit_segs = get_sb(segment_count_sit);
941 nat_segs = get_sb(segment_count_nat);
943 fsmeta += get_cp(rsvd_segment_count);
944 fsmeta += get_sb(segment_count_ssa);
949 ovp_segments = get_cp(overprov_segment_count);
950 reserved_segments = get_cp(rsvd_segment_count);
952 if (fsmeta < F2FS_MIN_SEGMENT || ovp_segments == 0 ||
953 reserved_segments == 0) {
954 MSG(0, "\tWrong layout: check mkfs.f2fs version\n");
958 user_block_count = get_cp(user_block_count);
959 segment_count_main = get_sb(segment_count_main);
960 log_blocks_per_seg = get_sb(log_blocks_per_seg);
961 if (!user_block_count || user_block_count >=
962 segment_count_main << log_blocks_per_seg) {
963 MSG(0, "\tWrong user_block_count(%u)\n", user_block_count);
967 main_segs = get_sb(segment_count_main);
968 blocks_per_seg = sbi->blocks_per_seg;
970 for (i = 0; i < NR_CURSEG_NODE_TYPE; i++) {
971 if (get_cp(cur_node_segno[i]) >= main_segs ||
972 get_cp(cur_node_blkoff[i]) >= blocks_per_seg)
975 for (i = 0; i < NR_CURSEG_DATA_TYPE; i++) {
976 if (get_cp(cur_data_segno[i]) >= main_segs ||
977 get_cp(cur_data_blkoff[i]) >= blocks_per_seg)
981 sit_bitmap_size = get_cp(sit_ver_bitmap_bytesize);
982 nat_bitmap_size = get_cp(nat_ver_bitmap_bytesize);
984 if (sit_bitmap_size != ((sit_segs / 2) << log_blocks_per_seg) / 8 ||
985 nat_bitmap_size != ((nat_segs / 2) << log_blocks_per_seg) / 8) {
986 MSG(0, "\tWrong bitmap size: sit(%u), nat(%u)\n",
987 sit_bitmap_size, nat_bitmap_size);
991 cp_pack_start_sum = __start_sum_addr(sbi);
992 cp_payload = __cp_payload(sbi);
993 if (cp_pack_start_sum < cp_payload + 1 ||
994 cp_pack_start_sum > blocks_per_seg - 1 -
996 MSG(0, "\tWrong cp_pack_start_sum(%u) or cp_payload(%u)\n",
997 cp_pack_start_sum, cp_payload);
998 if ((get_sb(feature) & F2FS_FEATURE_SB_CHKSUM))
1000 set_sb(cp_payload, cp_pack_start_sum - 1);
1001 update_superblock(sb, SB_MASK_ALL);
1007 pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start, int *pack)
1009 struct f2fs_nm_info *nm_i = NM_I(sbi);
1014 block_off = NAT_BLOCK_OFFSET(start);
1015 seg_off = block_off >> sbi->log_blocks_per_seg;
1017 block_addr = (pgoff_t)(nm_i->nat_blkaddr +
1018 (seg_off << sbi->log_blocks_per_seg << 1) +
1019 (block_off & ((1 << sbi->log_blocks_per_seg) -1)));
1023 if (f2fs_test_bit(block_off, nm_i->nat_bitmap)) {
1024 block_addr += sbi->blocks_per_seg;
1032 static int f2fs_init_nid_bitmap(struct f2fs_sb_info *sbi)
1034 struct f2fs_nm_info *nm_i = NM_I(sbi);
1035 int nid_bitmap_size = (nm_i->max_nid + BITS_PER_BYTE - 1) / BITS_PER_BYTE;
1036 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
1037 struct f2fs_summary_block *sum = curseg->sum_blk;
1038 struct f2fs_journal *journal = &sum->journal;
1039 struct f2fs_nat_block *nat_block;
1044 if (!(c.func == SLOAD || c.func == FSCK))
1047 nm_i->nid_bitmap = (char *)calloc(nid_bitmap_size, 1);
1048 if (!nm_i->nid_bitmap)
1051 /* arbitrarily set 0 bit */
1052 f2fs_set_bit(0, nm_i->nid_bitmap);
1054 nat_block = malloc(F2FS_BLKSIZE);
1056 free(nm_i->nid_bitmap);
1060 for (nid = 0; nid < nm_i->max_nid; nid++) {
1061 if (!(nid % NAT_ENTRY_PER_BLOCK)) {
1064 start_blk = current_nat_addr(sbi, nid, NULL);
1065 ret = dev_read_block(nat_block, start_blk);
1069 if (nat_block->entries[nid % NAT_ENTRY_PER_BLOCK].block_addr)
1070 f2fs_set_bit(nid, nm_i->nid_bitmap);
1073 if (nats_in_cursum(journal) > NAT_JOURNAL_ENTRIES) {
1074 MSG(0, "\tError: f2fs_init_nid_bitmap truncate n_nats(%u) to "
1075 "NAT_JOURNAL_ENTRIES(%lu)\n",
1076 nats_in_cursum(journal), NAT_JOURNAL_ENTRIES);
1077 journal->n_nats = cpu_to_le16(NAT_JOURNAL_ENTRIES);
1081 for (i = 0; i < nats_in_cursum(journal); i++) {
1084 addr = le32_to_cpu(nat_in_journal(journal, i).block_addr);
1085 if (!IS_VALID_BLK_ADDR(sbi, addr)) {
1086 MSG(0, "\tError: f2fs_init_nid_bitmap: addr(%u) is invalid!!!\n", addr);
1087 journal->n_nats = cpu_to_le16(i);
1092 nid = le32_to_cpu(nid_in_journal(journal, i));
1093 if (!IS_VALID_NID(sbi, nid)) {
1094 MSG(0, "\tError: f2fs_init_nid_bitmap: nid(%u) is invalid!!!\n", nid);
1095 journal->n_nats = cpu_to_le16(i);
1099 if (addr != NULL_ADDR)
1100 f2fs_set_bit(nid, nm_i->nid_bitmap);
1106 u32 update_nat_bits_flags(struct f2fs_super_block *sb,
1107 struct f2fs_checkpoint *cp, u32 flags)
1109 u_int32_t nat_bits_bytes, nat_bits_blocks;
1111 nat_bits_bytes = get_sb(segment_count_nat) << 5;
1112 nat_bits_blocks = F2FS_BYTES_TO_BLK((nat_bits_bytes << 1) + 8 +
1114 if (get_cp(cp_pack_total_block_count) <=
1115 (1 << get_sb(log_blocks_per_seg)) - nat_bits_blocks)
1116 flags |= CP_NAT_BITS_FLAG;
1118 flags &= (~CP_NAT_BITS_FLAG);
1123 /* should call flush_journal_entries() bfore this */
1124 void write_nat_bits(struct f2fs_sb_info *sbi,
1125 struct f2fs_super_block *sb, struct f2fs_checkpoint *cp, int set)
1127 struct f2fs_nm_info *nm_i = NM_I(sbi);
1128 u_int32_t nat_blocks = get_sb(segment_count_nat) <<
1129 (get_sb(log_blocks_per_seg) - 1);
1130 u_int32_t nat_bits_bytes = nat_blocks >> 3;
1131 u_int32_t nat_bits_blocks = F2FS_BYTES_TO_BLK((nat_bits_bytes << 1) +
1132 8 + F2FS_BLKSIZE - 1);
1133 unsigned char *nat_bits, *full_nat_bits, *empty_nat_bits;
1134 struct f2fs_nat_block *nat_block;
1139 nat_bits = calloc(F2FS_BLKSIZE, nat_bits_blocks);
1142 nat_block = malloc(F2FS_BLKSIZE);
1145 full_nat_bits = nat_bits + 8;
1146 empty_nat_bits = full_nat_bits + nat_bits_bytes;
1148 memset(full_nat_bits, 0, nat_bits_bytes);
1149 memset(empty_nat_bits, 0, nat_bits_bytes);
1151 for (i = 0; i < nat_blocks; i++) {
1152 int seg_off = i >> get_sb(log_blocks_per_seg);
1155 blkaddr = (pgoff_t)(get_sb(nat_blkaddr) +
1156 (seg_off << get_sb(log_blocks_per_seg) << 1) +
1157 (i & ((1 << get_sb(log_blocks_per_seg)) - 1)));
1160 * Should consider new nat_blocks is larger than old
1161 * nm_i->nat_blocks, since nm_i->nat_bitmap is based on
1164 if (i < nm_i->nat_blocks && f2fs_test_bit(i, nm_i->nat_bitmap))
1165 blkaddr += (1 << get_sb(log_blocks_per_seg));
1167 ret = dev_read_block(nat_block, blkaddr);
1170 for (j = 0; j < NAT_ENTRY_PER_BLOCK; j++) {
1171 if ((i == 0 && j == 0) ||
1172 nat_block->entries[j].block_addr != NULL_ADDR)
1176 test_and_set_bit_le(i, empty_nat_bits);
1177 else if (valid == NAT_ENTRY_PER_BLOCK)
1178 test_and_set_bit_le(i, full_nat_bits);
1180 *(__le64 *)nat_bits = get_cp_crc(cp);
1183 blkaddr = get_sb(segment0_blkaddr) + (set <<
1184 get_sb(log_blocks_per_seg)) - nat_bits_blocks;
1186 DBG(1, "\tWriting NAT bits pages, at offset 0x%08x\n", blkaddr);
1188 for (i = 0; i < nat_bits_blocks; i++) {
1189 if (dev_write_block(nat_bits + i * F2FS_BLKSIZE, blkaddr + i))
1190 ASSERT_MSG("\tError: write NAT bits to disk!!!\n");
1192 MSG(0, "Info: Write valid nat_bits in checkpoint\n");
1197 static int check_nat_bits(struct f2fs_sb_info *sbi,
1198 struct f2fs_super_block *sb, struct f2fs_checkpoint *cp)
1200 struct f2fs_nm_info *nm_i = NM_I(sbi);
1201 u_int32_t nat_blocks = get_sb(segment_count_nat) <<
1202 (get_sb(log_blocks_per_seg) - 1);
1203 u_int32_t nat_bits_bytes = nat_blocks >> 3;
1204 u_int32_t nat_bits_blocks = F2FS_BYTES_TO_BLK((nat_bits_bytes << 1) +
1205 8 + F2FS_BLKSIZE - 1);
1206 unsigned char *nat_bits, *full_nat_bits, *empty_nat_bits;
1207 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
1208 struct f2fs_journal *journal = &curseg->sum_blk->journal;
1213 nat_bits = calloc(F2FS_BLKSIZE, nat_bits_blocks);
1216 full_nat_bits = nat_bits + 8;
1217 empty_nat_bits = full_nat_bits + nat_bits_bytes;
1219 blkaddr = get_sb(segment0_blkaddr) + (sbi->cur_cp <<
1220 get_sb(log_blocks_per_seg)) - nat_bits_blocks;
1222 for (i = 0; i < nat_bits_blocks; i++) {
1223 if (dev_read_block(nat_bits + i * F2FS_BLKSIZE, blkaddr + i))
1224 ASSERT_MSG("\tError: read NAT bits to disk!!!\n");
1227 if (*(__le64 *)nat_bits != get_cp_crc(cp) || nats_in_cursum(journal)) {
1229 * if there is a journal, f2fs was not shutdown cleanly. Let's
1230 * flush them with nat_bits.
1234 /* Otherwise, kernel will disable nat_bits */
1238 for (i = 0; i < nat_blocks; i++) {
1239 u_int32_t start_nid = i * NAT_ENTRY_PER_BLOCK;
1240 u_int32_t valid = 0;
1241 int empty = test_bit_le(i, empty_nat_bits);
1242 int full = test_bit_le(i, full_nat_bits);
1244 for (j = 0; j < NAT_ENTRY_PER_BLOCK; j++) {
1245 if (f2fs_test_bit(start_nid + j, nm_i->nid_bitmap))
1249 if (!empty || full) {
1253 } else if (valid == NAT_ENTRY_PER_BLOCK) {
1254 if (empty || !full) {
1259 if (empty || full) {
1268 MSG(0, "Info: Checked valid nat_bits in checkpoint\n");
1271 MSG(0, "Info: Corrupted valid nat_bits in checkpoint\n");
1276 int init_node_manager(struct f2fs_sb_info *sbi)
1278 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
1279 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1280 struct f2fs_nm_info *nm_i = NM_I(sbi);
1281 unsigned char *version_bitmap;
1282 unsigned int nat_segs;
1284 nm_i->nat_blkaddr = get_sb(nat_blkaddr);
1286 /* segment_count_nat includes pair segment so divide to 2. */
1287 nat_segs = get_sb(segment_count_nat) >> 1;
1288 nm_i->nat_blocks = nat_segs << get_sb(log_blocks_per_seg);
1289 nm_i->max_nid = NAT_ENTRY_PER_BLOCK * nm_i->nat_blocks;
1292 nm_i->init_scan_nid = get_cp(next_free_nid);
1293 nm_i->next_scan_nid = get_cp(next_free_nid);
1295 nm_i->bitmap_size = __bitmap_size(sbi, NAT_BITMAP);
1297 nm_i->nat_bitmap = malloc(nm_i->bitmap_size);
1298 if (!nm_i->nat_bitmap)
1300 version_bitmap = __bitmap_ptr(sbi, NAT_BITMAP);
1301 if (!version_bitmap)
1304 /* copy version bitmap */
1305 memcpy(nm_i->nat_bitmap, version_bitmap, nm_i->bitmap_size);
1306 return f2fs_init_nid_bitmap(sbi);
1309 int build_node_manager(struct f2fs_sb_info *sbi)
1312 sbi->nm_info = malloc(sizeof(struct f2fs_nm_info));
1316 err = init_node_manager(sbi);
1323 int build_sit_info(struct f2fs_sb_info *sbi)
1325 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
1326 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1327 struct sit_info *sit_i;
1328 unsigned int sit_segs, start;
1329 char *src_bitmap, *dst_bitmap;
1330 unsigned int bitmap_size;
1332 sit_i = malloc(sizeof(struct sit_info));
1334 MSG(1, "\tError: Malloc failed for build_sit_info!\n");
1338 SM_I(sbi)->sit_info = sit_i;
1340 sit_i->sentries = calloc(TOTAL_SEGS(sbi) * sizeof(struct seg_entry), 1);
1341 if (!sit_i->sentries) {
1342 MSG(1, "\tError: Calloc failed for build_sit_info!\n");
1346 for (start = 0; start < TOTAL_SEGS(sbi); start++) {
1347 sit_i->sentries[start].cur_valid_map
1348 = calloc(SIT_VBLOCK_MAP_SIZE, 1);
1349 if (!sit_i->sentries[start].cur_valid_map) {
1350 MSG(1, "\tError: Calloc failed for build_sit_info!!\n");
1351 goto free_validity_maps;
1355 sit_segs = get_sb(segment_count_sit) >> 1;
1356 bitmap_size = __bitmap_size(sbi, SIT_BITMAP);
1357 src_bitmap = __bitmap_ptr(sbi, SIT_BITMAP);
1359 dst_bitmap = malloc(bitmap_size);
1361 MSG(1, "\tError: Malloc failed for build_sit_info!!\n");
1362 goto free_validity_maps;
1365 memcpy(dst_bitmap, src_bitmap, bitmap_size);
1367 sit_i->sit_base_addr = get_sb(sit_blkaddr);
1368 sit_i->sit_blocks = sit_segs << sbi->log_blocks_per_seg;
1369 sit_i->written_valid_blocks = get_cp(valid_block_count);
1370 sit_i->sit_bitmap = dst_bitmap;
1371 sit_i->bitmap_size = bitmap_size;
1372 sit_i->dirty_sentries = 0;
1373 sit_i->sents_per_block = SIT_ENTRY_PER_BLOCK;
1374 sit_i->elapsed_time = get_cp(elapsed_time);
1378 for (--start ; start >= 0; --start)
1379 free(sit_i->sentries[start].cur_valid_map);
1380 free(sit_i->sentries);
1388 void reset_curseg(struct f2fs_sb_info *sbi, int type)
1390 struct curseg_info *curseg = CURSEG_I(sbi, type);
1391 struct summary_footer *sum_footer;
1392 struct seg_entry *se;
1394 sum_footer = &(curseg->sum_blk->footer);
1395 memset(sum_footer, 0, sizeof(struct summary_footer));
1396 if (IS_DATASEG(type))
1397 SET_SUM_TYPE(sum_footer, SUM_TYPE_DATA);
1398 if (IS_NODESEG(type))
1399 SET_SUM_TYPE(sum_footer, SUM_TYPE_NODE);
1400 se = get_seg_entry(sbi, curseg->segno);
1405 static void read_compacted_summaries(struct f2fs_sb_info *sbi)
1407 struct curseg_info *curseg;
1408 unsigned int i, j, offset;
1413 start = start_sum_block(sbi);
1415 kaddr = (char *)malloc(PAGE_SIZE);
1418 ret = dev_read_block(kaddr, start++);
1421 curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
1422 memcpy(&curseg->sum_blk->journal.n_nats, kaddr, SUM_JOURNAL_SIZE);
1424 curseg = CURSEG_I(sbi, CURSEG_COLD_DATA);
1425 memcpy(&curseg->sum_blk->journal.n_sits, kaddr + SUM_JOURNAL_SIZE,
1428 offset = 2 * SUM_JOURNAL_SIZE;
1429 for (i = CURSEG_HOT_DATA; i <= CURSEG_COLD_DATA; i++) {
1430 unsigned short blk_off;
1431 struct curseg_info *curseg = CURSEG_I(sbi, i);
1433 reset_curseg(sbi, i);
1435 if (curseg->alloc_type == SSR)
1436 blk_off = sbi->blocks_per_seg;
1438 blk_off = curseg->next_blkoff;
1440 ASSERT(blk_off <= ENTRIES_IN_SUM);
1442 for (j = 0; j < blk_off; j++) {
1443 struct f2fs_summary *s;
1444 s = (struct f2fs_summary *)(kaddr + offset);
1445 curseg->sum_blk->entries[j] = *s;
1446 offset += SUMMARY_SIZE;
1447 if (offset + SUMMARY_SIZE <=
1448 PAGE_CACHE_SIZE - SUM_FOOTER_SIZE)
1450 memset(kaddr, 0, PAGE_SIZE);
1451 ret = dev_read_block(kaddr, start++);
1459 static void restore_node_summary(struct f2fs_sb_info *sbi,
1460 unsigned int segno, struct f2fs_summary_block *sum_blk)
1462 struct f2fs_node *node_blk;
1463 struct f2fs_summary *sum_entry;
1468 node_blk = malloc(F2FS_BLKSIZE);
1471 /* scan the node segment */
1472 addr = START_BLOCK(sbi, segno);
1473 sum_entry = &sum_blk->entries[0];
1475 for (i = 0; i < sbi->blocks_per_seg; i++, sum_entry++) {
1476 ret = dev_read_block(node_blk, addr);
1478 sum_entry->nid = node_blk->footer.nid;
1484 static void read_normal_summaries(struct f2fs_sb_info *sbi, int type)
1486 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1487 struct f2fs_summary_block *sum_blk;
1488 struct curseg_info *curseg;
1489 unsigned int segno = 0;
1490 block_t blk_addr = 0;
1493 if (IS_DATASEG(type)) {
1494 segno = get_cp(cur_data_segno[type]);
1495 if (is_set_ckpt_flags(cp, CP_UMOUNT_FLAG))
1496 blk_addr = sum_blk_addr(sbi, NR_CURSEG_TYPE, type);
1498 blk_addr = sum_blk_addr(sbi, NR_CURSEG_DATA_TYPE, type);
1500 segno = get_cp(cur_node_segno[type - CURSEG_HOT_NODE]);
1501 if (is_set_ckpt_flags(cp, CP_UMOUNT_FLAG))
1502 blk_addr = sum_blk_addr(sbi, NR_CURSEG_NODE_TYPE,
1503 type - CURSEG_HOT_NODE);
1505 blk_addr = GET_SUM_BLKADDR(sbi, segno);
1508 sum_blk = (struct f2fs_summary_block *)malloc(PAGE_SIZE);
1511 ret = dev_read_block(sum_blk, blk_addr);
1514 if (IS_NODESEG(type) && !is_set_ckpt_flags(cp, CP_UMOUNT_FLAG))
1515 restore_node_summary(sbi, segno, sum_blk);
1517 curseg = CURSEG_I(sbi, type);
1518 memcpy(curseg->sum_blk, sum_blk, PAGE_CACHE_SIZE);
1519 reset_curseg(sbi, type);
1523 void update_sum_entry(struct f2fs_sb_info *sbi, block_t blk_addr,
1524 struct f2fs_summary *sum)
1526 struct f2fs_summary_block *sum_blk;
1529 struct seg_entry *se;
1531 segno = GET_SEGNO(sbi, blk_addr);
1532 offset = OFFSET_IN_SEG(sbi, blk_addr);
1534 se = get_seg_entry(sbi, segno);
1536 sum_blk = get_sum_block(sbi, segno, &type);
1537 memcpy(&sum_blk->entries[offset], sum, sizeof(*sum));
1538 sum_blk->footer.entry_type = IS_NODESEG(se->type) ? SUM_TYPE_NODE :
1541 /* write SSA all the time */
1542 ret = dev_write_block(sum_blk, GET_SUM_BLKADDR(sbi, segno));
1545 if (type == SEG_TYPE_NODE || type == SEG_TYPE_DATA ||
1546 type == SEG_TYPE_MAX)
1550 static void restore_curseg_summaries(struct f2fs_sb_info *sbi)
1552 int type = CURSEG_HOT_DATA;
1554 if (is_set_ckpt_flags(F2FS_CKPT(sbi), CP_COMPACT_SUM_FLAG)) {
1555 read_compacted_summaries(sbi);
1556 type = CURSEG_HOT_NODE;
1559 for (; type <= CURSEG_COLD_NODE; type++)
1560 read_normal_summaries(sbi, type);
1563 static int build_curseg(struct f2fs_sb_info *sbi)
1565 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1566 struct curseg_info *array;
1567 unsigned short blk_off;
1571 array = malloc(sizeof(*array) * NR_CURSEG_TYPE);
1573 MSG(1, "\tError: Malloc failed for build_curseg!\n");
1577 SM_I(sbi)->curseg_array = array;
1579 for (i = 0; i < NR_CURSEG_TYPE; i++) {
1580 array[i].sum_blk = malloc(PAGE_CACHE_SIZE);
1581 if (!array[i].sum_blk) {
1582 MSG(1, "\tError: Malloc failed for build_curseg!!\n");
1586 if (i <= CURSEG_COLD_DATA) {
1587 blk_off = get_cp(cur_data_blkoff[i]);
1588 segno = get_cp(cur_data_segno[i]);
1590 if (i > CURSEG_COLD_DATA) {
1591 blk_off = get_cp(cur_node_blkoff[i - CURSEG_HOT_NODE]);
1592 segno = get_cp(cur_node_segno[i - CURSEG_HOT_NODE]);
1594 ASSERT(segno < TOTAL_SEGS(sbi));
1595 ASSERT(blk_off < DEFAULT_BLOCKS_PER_SEGMENT);
1597 array[i].segno = segno;
1598 array[i].zone = GET_ZONENO_FROM_SEGNO(sbi, segno);
1599 array[i].next_segno = NULL_SEGNO;
1600 array[i].next_blkoff = blk_off;
1601 array[i].alloc_type = cp->alloc_type[i];
1603 restore_curseg_summaries(sbi);
1607 for(--i ; i >=0; --i)
1608 free(array[i].sum_blk);
1614 static inline void check_seg_range(struct f2fs_sb_info *sbi, unsigned int segno)
1616 unsigned int end_segno = SM_I(sbi)->segment_count - 1;
1617 ASSERT(segno <= end_segno);
1620 void get_current_sit_page(struct f2fs_sb_info *sbi,
1621 unsigned int segno, struct f2fs_sit_block *sit_blk)
1623 struct sit_info *sit_i = SIT_I(sbi);
1624 unsigned int offset = SIT_BLOCK_OFFSET(sit_i, segno);
1625 block_t blk_addr = sit_i->sit_base_addr + offset;
1628 check_seg_range(sbi, segno);
1630 /* calculate sit block address */
1631 if (f2fs_test_bit(offset, sit_i->sit_bitmap))
1632 blk_addr += sit_i->sit_blocks;
1634 ret = dev_read_block(sit_blk, blk_addr);
1638 void rewrite_current_sit_page(struct f2fs_sb_info *sbi,
1639 unsigned int segno, struct f2fs_sit_block *sit_blk)
1641 struct sit_info *sit_i = SIT_I(sbi);
1642 unsigned int offset = SIT_BLOCK_OFFSET(sit_i, segno);
1643 block_t blk_addr = sit_i->sit_base_addr + offset;
1646 /* calculate sit block address */
1647 if (f2fs_test_bit(offset, sit_i->sit_bitmap))
1648 blk_addr += sit_i->sit_blocks;
1650 ret = dev_write_block(sit_blk, blk_addr);
1654 void check_block_count(struct f2fs_sb_info *sbi,
1655 unsigned int segno, struct f2fs_sit_entry *raw_sit)
1657 struct f2fs_sm_info *sm_info = SM_I(sbi);
1658 unsigned int end_segno = sm_info->segment_count - 1;
1659 int valid_blocks = 0;
1662 /* check segment usage */
1663 if (GET_SIT_VBLOCKS(raw_sit) > sbi->blocks_per_seg)
1664 ASSERT_MSG("Invalid SIT vblocks: segno=0x%x, %u",
1665 segno, GET_SIT_VBLOCKS(raw_sit));
1667 /* check boundary of a given segment number */
1668 if (segno > end_segno)
1669 ASSERT_MSG("Invalid SEGNO: 0x%x", segno);
1671 /* check bitmap with valid block count */
1672 for (i = 0; i < SIT_VBLOCK_MAP_SIZE; i++)
1673 valid_blocks += get_bits_in_byte(raw_sit->valid_map[i]);
1675 if (GET_SIT_VBLOCKS(raw_sit) != valid_blocks)
1676 ASSERT_MSG("Wrong SIT valid blocks: segno=0x%x, %u vs. %u",
1677 segno, GET_SIT_VBLOCKS(raw_sit), valid_blocks);
1679 if (GET_SIT_TYPE(raw_sit) >= NO_CHECK_TYPE)
1680 ASSERT_MSG("Wrong SIT type: segno=0x%x, %u",
1681 segno, GET_SIT_TYPE(raw_sit));
1684 void seg_info_from_raw_sit(struct seg_entry *se,
1685 struct f2fs_sit_entry *raw_sit)
1687 se->valid_blocks = GET_SIT_VBLOCKS(raw_sit);
1688 memcpy(se->cur_valid_map, raw_sit->valid_map, SIT_VBLOCK_MAP_SIZE);
1689 se->type = GET_SIT_TYPE(raw_sit);
1690 se->orig_type = GET_SIT_TYPE(raw_sit);
1691 se->mtime = le64_to_cpu(raw_sit->mtime);
1694 struct seg_entry *get_seg_entry(struct f2fs_sb_info *sbi,
1697 struct sit_info *sit_i = SIT_I(sbi);
1698 return &sit_i->sentries[segno];
1701 struct f2fs_summary_block *get_sum_block(struct f2fs_sb_info *sbi,
1702 unsigned int segno, int *ret_type)
1704 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1705 struct f2fs_summary_block *sum_blk;
1706 struct curseg_info *curseg;
1710 *ret_type= SEG_TYPE_MAX;
1712 ssa_blk = GET_SUM_BLKADDR(sbi, segno);
1713 for (type = 0; type < NR_CURSEG_NODE_TYPE; type++) {
1714 if (segno == get_cp(cur_node_segno[type])) {
1715 curseg = CURSEG_I(sbi, CURSEG_HOT_NODE + type);
1716 if (!IS_SUM_NODE_SEG(curseg->sum_blk->footer)) {
1717 ASSERT_MSG("segno [0x%x] indicates a data "
1718 "segment, but should be node",
1720 *ret_type = -SEG_TYPE_CUR_NODE;
1722 *ret_type = SEG_TYPE_CUR_NODE;
1724 return curseg->sum_blk;
1728 for (type = 0; type < NR_CURSEG_DATA_TYPE; type++) {
1729 if (segno == get_cp(cur_data_segno[type])) {
1730 curseg = CURSEG_I(sbi, type);
1731 if (IS_SUM_NODE_SEG(curseg->sum_blk->footer)) {
1732 ASSERT_MSG("segno [0x%x] indicates a node "
1733 "segment, but should be data",
1735 *ret_type = -SEG_TYPE_CUR_DATA;
1737 *ret_type = SEG_TYPE_CUR_DATA;
1739 return curseg->sum_blk;
1743 sum_blk = calloc(BLOCK_SZ, 1);
1746 ret = dev_read_block(sum_blk, ssa_blk);
1749 if (IS_SUM_NODE_SEG(sum_blk->footer))
1750 *ret_type = SEG_TYPE_NODE;
1751 else if (IS_SUM_DATA_SEG(sum_blk->footer))
1752 *ret_type = SEG_TYPE_DATA;
1757 int get_sum_entry(struct f2fs_sb_info *sbi, u32 blk_addr,
1758 struct f2fs_summary *sum_entry)
1760 struct f2fs_summary_block *sum_blk;
1764 segno = GET_SEGNO(sbi, blk_addr);
1765 offset = OFFSET_IN_SEG(sbi, blk_addr);
1767 sum_blk = get_sum_block(sbi, segno, &type);
1768 memcpy(sum_entry, &(sum_blk->entries[offset]),
1769 sizeof(struct f2fs_summary));
1770 if (type == SEG_TYPE_NODE || type == SEG_TYPE_DATA ||
1771 type == SEG_TYPE_MAX)
1776 static void get_nat_entry(struct f2fs_sb_info *sbi, nid_t nid,
1777 struct f2fs_nat_entry *raw_nat)
1779 struct f2fs_nat_block *nat_block;
1784 if (lookup_nat_in_journal(sbi, nid, raw_nat) >= 0)
1787 nat_block = (struct f2fs_nat_block *)calloc(BLOCK_SZ, 1);
1790 entry_off = nid % NAT_ENTRY_PER_BLOCK;
1791 block_addr = current_nat_addr(sbi, nid, NULL);
1793 ret = dev_read_block(nat_block, block_addr);
1796 memcpy(raw_nat, &nat_block->entries[entry_off],
1797 sizeof(struct f2fs_nat_entry));
1801 void update_data_blkaddr(struct f2fs_sb_info *sbi, nid_t nid,
1802 u16 ofs_in_node, block_t newaddr)
1804 struct f2fs_node *node_blk = NULL;
1805 struct node_info ni;
1806 block_t oldaddr, startaddr, endaddr;
1809 node_blk = (struct f2fs_node *)calloc(BLOCK_SZ, 1);
1812 get_node_info(sbi, nid, &ni);
1814 /* read node_block */
1815 ret = dev_read_block(node_blk, ni.blk_addr);
1818 /* check its block address */
1819 if (node_blk->footer.nid == node_blk->footer.ino) {
1820 int ofs = get_extra_isize(node_blk);
1822 oldaddr = le32_to_cpu(node_blk->i.i_addr[ofs + ofs_in_node]);
1823 node_blk->i.i_addr[ofs + ofs_in_node] = cpu_to_le32(newaddr);
1824 ret = write_inode(node_blk, ni.blk_addr);
1827 oldaddr = le32_to_cpu(node_blk->dn.addr[ofs_in_node]);
1828 node_blk->dn.addr[ofs_in_node] = cpu_to_le32(newaddr);
1829 ret = dev_write_block(node_blk, ni.blk_addr);
1833 /* check extent cache entry */
1834 if (node_blk->footer.nid != node_blk->footer.ino) {
1835 get_node_info(sbi, le32_to_cpu(node_blk->footer.ino), &ni);
1837 /* read inode block */
1838 ret = dev_read_block(node_blk, ni.blk_addr);
1842 startaddr = le32_to_cpu(node_blk->i.i_ext.blk_addr);
1843 endaddr = startaddr + le32_to_cpu(node_blk->i.i_ext.len);
1844 if (oldaddr >= startaddr && oldaddr < endaddr) {
1845 node_blk->i.i_ext.len = 0;
1847 /* update inode block */
1848 ASSERT(write_inode(node_blk, ni.blk_addr) >= 0);
1853 void update_nat_blkaddr(struct f2fs_sb_info *sbi, nid_t ino,
1854 nid_t nid, block_t newaddr)
1856 struct f2fs_nat_block *nat_block;
1861 nat_block = (struct f2fs_nat_block *)calloc(BLOCK_SZ, 1);
1864 entry_off = nid % NAT_ENTRY_PER_BLOCK;
1865 block_addr = current_nat_addr(sbi, nid, NULL);
1867 ret = dev_read_block(nat_block, block_addr);
1871 nat_block->entries[entry_off].ino = cpu_to_le32(ino);
1872 nat_block->entries[entry_off].block_addr = cpu_to_le32(newaddr);
1874 F2FS_FSCK(sbi)->entries[nid] = nat_block->entries[entry_off];
1876 ret = dev_write_block(nat_block, block_addr);
1881 void get_node_info(struct f2fs_sb_info *sbi, nid_t nid, struct node_info *ni)
1883 struct f2fs_nat_entry raw_nat;
1886 if (c.func == FSCK) {
1887 node_info_from_raw_nat(ni, &(F2FS_FSCK(sbi)->entries[nid]));
1890 /* nat entry is not cached, read it */
1893 get_nat_entry(sbi, nid, &raw_nat);
1894 node_info_from_raw_nat(ni, &raw_nat);
1897 static int build_sit_entries(struct f2fs_sb_info *sbi)
1899 struct sit_info *sit_i = SIT_I(sbi);
1900 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_COLD_DATA);
1901 struct f2fs_journal *journal = &curseg->sum_blk->journal;
1902 struct f2fs_sit_block *sit_blk;
1903 struct seg_entry *se;
1904 struct f2fs_sit_entry sit;
1905 unsigned int i, segno;
1907 sit_blk = calloc(BLOCK_SZ, 1);
1909 MSG(1, "\tError: Calloc failed for build_sit_entries!\n");
1913 for (segno = 0; segno < TOTAL_SEGS(sbi); segno++) {
1914 se = &sit_i->sentries[segno];
1916 get_current_sit_page(sbi, segno, sit_blk);
1917 sit = sit_blk->entries[SIT_ENTRY_OFFSET(sit_i, segno)];
1919 check_block_count(sbi, segno, &sit);
1920 seg_info_from_raw_sit(se, &sit);
1925 if (sits_in_cursum(journal) > SIT_JOURNAL_ENTRIES) {
1926 MSG(0, "\tError: build_sit_entries truncate n_sits(%u) to "
1927 "SIT_JOURNAL_ENTRIES(%lu)\n",
1928 sits_in_cursum(journal), SIT_JOURNAL_ENTRIES);
1929 journal->n_sits = cpu_to_le16(SIT_JOURNAL_ENTRIES);
1933 for (i = 0; i < sits_in_cursum(journal); i++) {
1934 segno = le32_to_cpu(segno_in_journal(journal, i));
1936 if (segno >= TOTAL_SEGS(sbi)) {
1937 MSG(0, "\tError: build_sit_entries: segno(%u) is invalid!!!\n", segno);
1938 journal->n_sits = cpu_to_le16(i);
1943 se = &sit_i->sentries[segno];
1944 sit = sit_in_journal(journal, i);
1946 check_block_count(sbi, segno, &sit);
1947 seg_info_from_raw_sit(se, &sit);
1952 static int build_segment_manager(struct f2fs_sb_info *sbi)
1954 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
1955 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
1956 struct f2fs_sm_info *sm_info;
1958 sm_info = malloc(sizeof(struct f2fs_sm_info));
1960 MSG(1, "\tError: Malloc failed for build_segment_manager!\n");
1965 sbi->sm_info = sm_info;
1966 sm_info->seg0_blkaddr = get_sb(segment0_blkaddr);
1967 sm_info->main_blkaddr = get_sb(main_blkaddr);
1968 sm_info->segment_count = get_sb(segment_count);
1969 sm_info->reserved_segments = get_cp(rsvd_segment_count);
1970 sm_info->ovp_segments = get_cp(overprov_segment_count);
1971 sm_info->main_segments = get_sb(segment_count_main);
1972 sm_info->ssa_blkaddr = get_sb(ssa_blkaddr);
1974 if (build_sit_info(sbi) || build_curseg(sbi) || build_sit_entries(sbi)) {
1982 void build_sit_area_bitmap(struct f2fs_sb_info *sbi)
1984 struct f2fs_fsck *fsck = F2FS_FSCK(sbi);
1985 struct f2fs_sm_info *sm_i = SM_I(sbi);
1986 unsigned int segno = 0;
1988 u32 sum_vblocks = 0;
1990 struct seg_entry *se;
1992 fsck->sit_area_bitmap_sz = sm_i->main_segments * SIT_VBLOCK_MAP_SIZE;
1993 fsck->sit_area_bitmap = calloc(1, fsck->sit_area_bitmap_sz);
1994 ASSERT(fsck->sit_area_bitmap);
1995 ptr = fsck->sit_area_bitmap;
1997 ASSERT(fsck->sit_area_bitmap_sz == fsck->main_area_bitmap_sz);
1999 for (segno = 0; segno < TOTAL_SEGS(sbi); segno++) {
2000 se = get_seg_entry(sbi, segno);
2002 memcpy(ptr, se->cur_valid_map, SIT_VBLOCK_MAP_SIZE);
2003 ptr += SIT_VBLOCK_MAP_SIZE;
2005 if (se->valid_blocks == 0x0) {
2006 if (le32_to_cpu(sbi->ckpt->cur_node_segno[0]) == segno ||
2007 le32_to_cpu(sbi->ckpt->cur_data_segno[0]) == segno ||
2008 le32_to_cpu(sbi->ckpt->cur_node_segno[1]) == segno ||
2009 le32_to_cpu(sbi->ckpt->cur_data_segno[1]) == segno ||
2010 le32_to_cpu(sbi->ckpt->cur_node_segno[2]) == segno ||
2011 le32_to_cpu(sbi->ckpt->cur_data_segno[2]) == segno) {
2017 sum_vblocks += se->valid_blocks;
2020 fsck->chk.sit_valid_blocks = sum_vblocks;
2021 fsck->chk.sit_free_segs = free_segs;
2023 DBG(1, "Blocks [0x%x : %d] Free Segs [0x%x : %d]\n\n",
2024 sum_vblocks, sum_vblocks,
2025 free_segs, free_segs);
2028 void rewrite_sit_area_bitmap(struct f2fs_sb_info *sbi)
2030 struct f2fs_fsck *fsck = F2FS_FSCK(sbi);
2031 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_COLD_DATA);
2032 struct sit_info *sit_i = SIT_I(sbi);
2033 struct f2fs_sit_block *sit_blk;
2034 unsigned int segno = 0;
2035 struct f2fs_summary_block *sum = curseg->sum_blk;
2038 sit_blk = calloc(BLOCK_SZ, 1);
2040 /* remove sit journal */
2041 sum->journal.n_sits = 0;
2043 ptr = fsck->main_area_bitmap;
2045 for (segno = 0; segno < TOTAL_SEGS(sbi); segno++) {
2046 struct f2fs_sit_entry *sit;
2047 struct seg_entry *se;
2048 u16 valid_blocks = 0;
2052 get_current_sit_page(sbi, segno, sit_blk);
2053 sit = &sit_blk->entries[SIT_ENTRY_OFFSET(sit_i, segno)];
2054 memcpy(sit->valid_map, ptr, SIT_VBLOCK_MAP_SIZE);
2056 /* update valid block count */
2057 for (i = 0; i < SIT_VBLOCK_MAP_SIZE; i++)
2058 valid_blocks += get_bits_in_byte(sit->valid_map[i]);
2060 se = get_seg_entry(sbi, segno);
2061 memcpy(se->cur_valid_map, ptr, SIT_VBLOCK_MAP_SIZE);
2062 se->valid_blocks = valid_blocks;
2064 if (type >= NO_CHECK_TYPE) {
2065 ASSERT_MSG("Invalide type and valid blocks=%x,%x",
2066 segno, valid_blocks);
2069 sit->vblocks = cpu_to_le16((type << SIT_VBLOCKS_SHIFT) |
2071 rewrite_current_sit_page(sbi, segno, sit_blk);
2073 ptr += SIT_VBLOCK_MAP_SIZE;
2079 static int flush_sit_journal_entries(struct f2fs_sb_info *sbi)
2081 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_COLD_DATA);
2082 struct f2fs_journal *journal = &curseg->sum_blk->journal;
2083 struct sit_info *sit_i = SIT_I(sbi);
2084 struct f2fs_sit_block *sit_blk;
2088 sit_blk = calloc(BLOCK_SZ, 1);
2090 for (i = 0; i < sits_in_cursum(journal); i++) {
2091 struct f2fs_sit_entry *sit;
2092 struct seg_entry *se;
2094 segno = segno_in_journal(journal, i);
2095 se = get_seg_entry(sbi, segno);
2097 get_current_sit_page(sbi, segno, sit_blk);
2098 sit = &sit_blk->entries[SIT_ENTRY_OFFSET(sit_i, segno)];
2100 memcpy(sit->valid_map, se->cur_valid_map, SIT_VBLOCK_MAP_SIZE);
2101 sit->vblocks = cpu_to_le16((se->type << SIT_VBLOCKS_SHIFT) |
2103 sit->mtime = cpu_to_le64(se->mtime);
2105 rewrite_current_sit_page(sbi, segno, sit_blk);
2109 journal->n_sits = 0;
2113 static int flush_nat_journal_entries(struct f2fs_sb_info *sbi)
2115 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
2116 struct f2fs_journal *journal = &curseg->sum_blk->journal;
2117 struct f2fs_nat_block *nat_block;
2124 nat_block = (struct f2fs_nat_block *)calloc(BLOCK_SZ, 1);
2127 if (i >= nats_in_cursum(journal)) {
2129 journal->n_nats = 0;
2133 nid = le32_to_cpu(nid_in_journal(journal, i));
2135 entry_off = nid % NAT_ENTRY_PER_BLOCK;
2136 block_addr = current_nat_addr(sbi, nid, NULL);
2138 ret = dev_read_block(nat_block, block_addr);
2141 memcpy(&nat_block->entries[entry_off], &nat_in_journal(journal, i),
2142 sizeof(struct f2fs_nat_entry));
2144 ret = dev_write_block(nat_block, block_addr);
2150 void flush_journal_entries(struct f2fs_sb_info *sbi)
2152 int n_nats = flush_nat_journal_entries(sbi);
2153 int n_sits = flush_sit_journal_entries(sbi);
2155 if (n_nats || n_sits)
2156 write_checkpoint(sbi);
2159 void flush_sit_entries(struct f2fs_sb_info *sbi)
2161 struct sit_info *sit_i = SIT_I(sbi);
2162 struct f2fs_sit_block *sit_blk;
2163 unsigned int segno = 0;
2165 sit_blk = calloc(BLOCK_SZ, 1);
2167 /* update free segments */
2168 for (segno = 0; segno < TOTAL_SEGS(sbi); segno++) {
2169 struct f2fs_sit_entry *sit;
2170 struct seg_entry *se;
2172 se = get_seg_entry(sbi, segno);
2177 get_current_sit_page(sbi, segno, sit_blk);
2178 sit = &sit_blk->entries[SIT_ENTRY_OFFSET(sit_i, segno)];
2179 memcpy(sit->valid_map, se->cur_valid_map, SIT_VBLOCK_MAP_SIZE);
2180 sit->vblocks = cpu_to_le16((se->type << SIT_VBLOCKS_SHIFT) |
2182 rewrite_current_sit_page(sbi, segno, sit_blk);
2188 int relocate_curseg_offset(struct f2fs_sb_info *sbi, int type)
2190 struct curseg_info *curseg = CURSEG_I(sbi, type);
2191 struct seg_entry *se = get_seg_entry(sbi, curseg->segno);
2194 for (i = 0; i < sbi->blocks_per_seg; i++) {
2195 if (!f2fs_test_bit(i, (const char *)se->cur_valid_map))
2199 if (i == sbi->blocks_per_seg)
2202 DBG(1, "Update curseg[%d].next_blkoff %u -> %u, alloc_type %s -> SSR\n",
2203 type, curseg->next_blkoff, i,
2204 curseg->alloc_type == LFS ? "LFS" : "SSR");
2206 curseg->next_blkoff = i;
2207 curseg->alloc_type = SSR;
2212 int find_next_free_block(struct f2fs_sb_info *sbi, u64 *to, int left, int type)
2214 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
2215 struct seg_entry *se;
2219 u64 end_blkaddr = (get_sb(segment_count_main) <<
2220 get_sb(log_blocks_per_seg)) + get_sb(main_blkaddr);
2224 if (get_free_segments(sbi) <= SM_I(sbi)->reserved_segments + 1)
2227 while (*to >= SM_I(sbi)->main_blkaddr && *to < end_blkaddr) {
2228 segno = GET_SEGNO(sbi, *to);
2229 offset = OFFSET_IN_SEG(sbi, *to);
2231 se = get_seg_entry(sbi, segno);
2233 if (se->valid_blocks == sbi->blocks_per_seg ||
2234 IS_CUR_SEGNO(sbi, segno)) {
2235 *to = left ? START_BLOCK(sbi, segno) - 1:
2236 START_BLOCK(sbi, segno + 1);
2240 if (se->valid_blocks == 0 && not_enough) {
2241 *to = left ? START_BLOCK(sbi, segno) - 1:
2242 START_BLOCK(sbi, segno + 1);
2246 if (se->valid_blocks == 0 && !(segno % sbi->segs_per_sec)) {
2247 struct seg_entry *se2;
2250 for (i = 1; i < sbi->segs_per_sec; i++) {
2251 se2 = get_seg_entry(sbi, segno + i);
2252 if (se2->valid_blocks)
2255 if (i == sbi->segs_per_sec)
2259 if (se->type == type &&
2260 !f2fs_test_bit(offset, (const char *)se->cur_valid_map))
2263 *to = left ? *to - 1: *to + 1;
2268 void move_curseg_info(struct f2fs_sb_info *sbi, u64 from, int left)
2272 /* update summary blocks having nullified journal entries */
2273 for (i = 0; i < NO_CHECK_TYPE; i++) {
2274 struct curseg_info *curseg = CURSEG_I(sbi, i);
2275 struct f2fs_summary_block buf;
2279 /* update original SSA too */
2280 ssa_blk = GET_SUM_BLKADDR(sbi, curseg->segno);
2281 ret = dev_write_block(curseg->sum_blk, ssa_blk);
2285 ret = find_next_free_block(sbi, &to, left, i);
2288 old_segno = curseg->segno;
2289 curseg->segno = GET_SEGNO(sbi, to);
2290 curseg->next_blkoff = OFFSET_IN_SEG(sbi, to);
2291 curseg->alloc_type = SSR;
2293 /* update new segno */
2294 ssa_blk = GET_SUM_BLKADDR(sbi, curseg->segno);
2295 ret = dev_read_block(&buf, ssa_blk);
2298 memcpy(curseg->sum_blk, &buf, SUM_ENTRIES_SIZE);
2300 /* update se->types */
2301 reset_curseg(sbi, i);
2303 DBG(1, "Move curseg[%d] %x -> %x after %"PRIx64"\n",
2304 i, old_segno, curseg->segno, from);
2308 void update_curseg_info(struct f2fs_sb_info *sbi, int type)
2310 if (!relocate_curseg_offset(sbi, type))
2312 move_curseg_info(sbi, SM_I(sbi)->main_blkaddr, 0);
2315 void zero_journal_entries(struct f2fs_sb_info *sbi)
2319 for (i = 0; i < NO_CHECK_TYPE; i++)
2320 CURSEG_I(sbi, i)->sum_blk->journal.n_nats = 0;
2323 void write_curseg_info(struct f2fs_sb_info *sbi)
2325 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
2328 for (i = 0; i < NO_CHECK_TYPE; i++) {
2329 cp->alloc_type[i] = CURSEG_I(sbi, i)->alloc_type;
2330 if (i < CURSEG_HOT_NODE) {
2331 set_cp(cur_data_segno[i], CURSEG_I(sbi, i)->segno);
2332 set_cp(cur_data_blkoff[i],
2333 CURSEG_I(sbi, i)->next_blkoff);
2335 int n = i - CURSEG_HOT_NODE;
2337 set_cp(cur_node_segno[n], CURSEG_I(sbi, i)->segno);
2338 set_cp(cur_node_blkoff[n],
2339 CURSEG_I(sbi, i)->next_blkoff);
2344 int lookup_nat_in_journal(struct f2fs_sb_info *sbi, u32 nid,
2345 struct f2fs_nat_entry *raw_nat)
2347 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
2348 struct f2fs_journal *journal = &curseg->sum_blk->journal;
2351 for (i = 0; i < nats_in_cursum(journal); i++) {
2352 if (le32_to_cpu(nid_in_journal(journal, i)) == nid) {
2353 memcpy(raw_nat, &nat_in_journal(journal, i),
2354 sizeof(struct f2fs_nat_entry));
2355 DBG(3, "==> Found nid [0x%x] in nat cache\n", nid);
2362 void nullify_nat_entry(struct f2fs_sb_info *sbi, u32 nid)
2364 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
2365 struct f2fs_journal *journal = &curseg->sum_blk->journal;
2366 struct f2fs_nat_block *nat_block;
2372 /* check in journal */
2373 for (i = 0; i < nats_in_cursum(journal); i++) {
2374 if (le32_to_cpu(nid_in_journal(journal, i)) == nid) {
2375 memset(&nat_in_journal(journal, i), 0,
2376 sizeof(struct f2fs_nat_entry));
2377 FIX_MSG("Remove nid [0x%x] in nat journal", nid);
2381 nat_block = (struct f2fs_nat_block *)calloc(BLOCK_SZ, 1);
2384 entry_off = nid % NAT_ENTRY_PER_BLOCK;
2385 block_addr = current_nat_addr(sbi, nid, NULL);
2387 ret = dev_read_block(nat_block, block_addr);
2390 if (nid == F2FS_NODE_INO(sbi) || nid == F2FS_META_INO(sbi)) {
2391 FIX_MSG("nid [0x%x] block_addr= 0x%x -> 0x1", nid,
2392 le32_to_cpu(nat_block->entries[entry_off].block_addr));
2393 nat_block->entries[entry_off].block_addr = cpu_to_le32(0x1);
2395 memset(&nat_block->entries[entry_off], 0,
2396 sizeof(struct f2fs_nat_entry));
2397 FIX_MSG("Remove nid [0x%x] in NAT", nid);
2400 ret = dev_write_block(nat_block, block_addr);
2405 void write_checkpoint(struct f2fs_sb_info *sbi)
2407 struct f2fs_checkpoint *cp = F2FS_CKPT(sbi);
2408 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
2409 block_t orphan_blks = 0;
2410 unsigned long long cp_blk_no;
2411 u32 flags = CP_UMOUNT_FLAG;
2415 if (is_set_ckpt_flags(cp, CP_ORPHAN_PRESENT_FLAG)) {
2416 orphan_blks = __start_sum_addr(sbi) - 1;
2417 flags |= CP_ORPHAN_PRESENT_FLAG;
2419 if (is_set_ckpt_flags(cp, CP_TRIMMED_FLAG))
2420 flags |= CP_TRIMMED_FLAG;
2421 if (is_set_ckpt_flags(cp, CP_DISABLED_FLAG))
2422 flags |= CP_DISABLED_FLAG;
2423 if (is_set_ckpt_flags(cp, CP_LARGE_NAT_BITMAP_FLAG)) {
2424 flags |= CP_LARGE_NAT_BITMAP_FLAG;
2425 set_cp(checksum_offset, CP_MIN_CHKSUM_OFFSET);
2427 set_cp(checksum_offset, CP_CHKSUM_OFFSET);
2430 set_cp(free_segment_count, get_free_segments(sbi));
2431 set_cp(valid_block_count, sbi->total_valid_block_count);
2432 set_cp(cp_pack_total_block_count, 8 + orphan_blks + get_sb(cp_payload));
2434 flags = update_nat_bits_flags(sb, cp, flags);
2435 set_cp(ckpt_flags, flags);
2437 crc = f2fs_checkpoint_chksum(cp);
2438 *((__le32 *)((unsigned char *)cp + get_cp(checksum_offset))) =
2441 cp_blk_no = get_sb(cp_blkaddr);
2442 if (sbi->cur_cp == 2)
2443 cp_blk_no += 1 << get_sb(log_blocks_per_seg);
2445 /* write the first cp */
2446 ret = dev_write_block(cp, cp_blk_no++);
2450 cp_blk_no += get_sb(cp_payload);
2451 /* skip orphan blocks */
2452 cp_blk_no += orphan_blks;
2454 /* update summary blocks having nullified journal entries */
2455 for (i = 0; i < NO_CHECK_TYPE; i++) {
2456 struct curseg_info *curseg = CURSEG_I(sbi, i);
2459 ret = dev_write_block(curseg->sum_blk, cp_blk_no++);
2462 /* update original SSA too */
2463 ssa_blk = GET_SUM_BLKADDR(sbi, curseg->segno);
2464 ret = dev_write_block(curseg->sum_blk, ssa_blk);
2468 /* Write nat bits */
2469 if (flags & CP_NAT_BITS_FLAG)
2470 write_nat_bits(sbi, sb, cp, sbi->cur_cp);
2472 /* in case of sudden power off */
2473 ret = f2fs_fsync_device();
2476 /* write the last cp */
2477 ret = dev_write_block(cp, cp_blk_no++);
2480 ret = f2fs_fsync_device();
2484 void build_nat_area_bitmap(struct f2fs_sb_info *sbi)
2486 struct curseg_info *curseg = CURSEG_I(sbi, CURSEG_HOT_DATA);
2487 struct f2fs_journal *journal = &curseg->sum_blk->journal;
2488 struct f2fs_fsck *fsck = F2FS_FSCK(sbi);
2489 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
2490 struct f2fs_nm_info *nm_i = NM_I(sbi);
2491 struct f2fs_nat_block *nat_block;
2492 struct node_info ni;
2493 u32 nid, nr_nat_blks;
2500 nat_block = (struct f2fs_nat_block *)calloc(BLOCK_SZ, 1);
2503 /* Alloc & build nat entry bitmap */
2504 nr_nat_blks = (get_sb(segment_count_nat) / 2) <<
2505 sbi->log_blocks_per_seg;
2507 fsck->nr_nat_entries = nr_nat_blks * NAT_ENTRY_PER_BLOCK;
2508 fsck->nat_area_bitmap_sz = (fsck->nr_nat_entries + 7) / 8;
2509 fsck->nat_area_bitmap = calloc(fsck->nat_area_bitmap_sz, 1);
2510 ASSERT(fsck->nat_area_bitmap);
2512 fsck->entries = calloc(sizeof(struct f2fs_nat_entry),
2513 fsck->nr_nat_entries);
2514 ASSERT(fsck->entries);
2516 for (block_off = 0; block_off < nr_nat_blks; block_off++) {
2518 seg_off = block_off >> sbi->log_blocks_per_seg;
2519 block_addr = (pgoff_t)(nm_i->nat_blkaddr +
2520 (seg_off << sbi->log_blocks_per_seg << 1) +
2521 (block_off & ((1 << sbi->log_blocks_per_seg) - 1)));
2523 if (f2fs_test_bit(block_off, nm_i->nat_bitmap))
2524 block_addr += sbi->blocks_per_seg;
2526 ret = dev_read_block(nat_block, block_addr);
2529 nid = block_off * NAT_ENTRY_PER_BLOCK;
2530 for (i = 0; i < NAT_ENTRY_PER_BLOCK; i++) {
2533 if ((nid + i) == F2FS_NODE_INO(sbi) ||
2534 (nid + i) == F2FS_META_INO(sbi)) {
2536 * block_addr of node/meta inode should be 0x1.
2537 * Set this bit, and fsck_verify will fix it.
2539 if (le32_to_cpu(nat_block->entries[i].block_addr) != 0x1) {
2540 ASSERT_MSG("\tError: ino[0x%x] block_addr[0x%x] is invalid\n",
2541 nid + i, le32_to_cpu(nat_block->entries[i].block_addr));
2542 f2fs_set_bit(nid + i, fsck->nat_area_bitmap);
2547 node_info_from_raw_nat(&ni, &nat_block->entries[i]);
2548 if (ni.blk_addr == 0x0)
2550 if (ni.ino == 0x0) {
2551 ASSERT_MSG("\tError: ino[0x%8x] or blk_addr[0x%16x]"
2552 " is invalid\n", ni.ino, ni.blk_addr);
2554 if (ni.ino == (nid + i)) {
2555 fsck->nat_valid_inode_cnt++;
2556 DBG(3, "ino[0x%8x] maybe is inode\n", ni.ino);
2560 * nat entry [0] must be null. If
2561 * it is corrupted, set its bit in
2562 * nat_area_bitmap, fsck_verify will
2565 ASSERT_MSG("Invalid nat entry[0]: "
2566 "blk_addr[0x%x]\n", ni.blk_addr);
2567 fsck->chk.valid_nat_entry_cnt--;
2570 DBG(3, "nid[0x%8x] addr[0x%16x] ino[0x%8x]\n",
2571 nid + i, ni.blk_addr, ni.ino);
2572 f2fs_set_bit(nid + i, fsck->nat_area_bitmap);
2573 fsck->chk.valid_nat_entry_cnt++;
2575 fsck->entries[nid + i] = nat_block->entries[i];
2579 /* Traverse nat journal, update the corresponding entries */
2580 for (i = 0; i < nats_in_cursum(journal); i++) {
2581 struct f2fs_nat_entry raw_nat;
2582 nid = le32_to_cpu(nid_in_journal(journal, i));
2585 DBG(3, "==> Found nid [0x%x] in nat cache, update it\n", nid);
2587 /* Clear the original bit and count */
2588 if (fsck->entries[nid].block_addr != 0x0) {
2589 fsck->chk.valid_nat_entry_cnt--;
2590 f2fs_clear_bit(nid, fsck->nat_area_bitmap);
2591 if (fsck->entries[nid].ino == nid)
2592 fsck->nat_valid_inode_cnt--;
2595 /* Use nat entries in journal */
2596 memcpy(&raw_nat, &nat_in_journal(journal, i),
2597 sizeof(struct f2fs_nat_entry));
2598 node_info_from_raw_nat(&ni, &raw_nat);
2599 if (ni.blk_addr != 0x0) {
2601 ASSERT_MSG("\tError: ino[0x%8x] or blk_addr[0x%16x]"
2602 " is invalid\n", ni.ino, ni.blk_addr);
2603 if (ni.ino == nid) {
2604 fsck->nat_valid_inode_cnt++;
2605 DBG(3, "ino[0x%8x] maybe is inode\n", ni.ino);
2607 f2fs_set_bit(nid, fsck->nat_area_bitmap);
2608 fsck->chk.valid_nat_entry_cnt++;
2609 DBG(3, "nid[0x%x] in nat cache\n", nid);
2611 fsck->entries[nid] = raw_nat;
2615 DBG(1, "valid nat entries (block_addr != 0x0) [0x%8x : %u]\n",
2616 fsck->chk.valid_nat_entry_cnt,
2617 fsck->chk.valid_nat_entry_cnt);
2620 static int check_sector_size(struct f2fs_super_block *sb)
2622 u_int32_t log_sectorsize, log_sectors_per_block;
2624 log_sectorsize = log_base_2(c.sector_size);
2625 log_sectors_per_block = log_base_2(c.sectors_per_blk);
2627 if (log_sectorsize == get_sb(log_sectorsize) &&
2628 log_sectors_per_block == get_sb(log_sectors_per_block))
2631 set_sb(log_sectorsize, log_sectorsize);
2632 set_sb(log_sectors_per_block, log_sectors_per_block);
2634 update_superblock(sb, SB_MASK_ALL);
2638 static void tune_sb_features(struct f2fs_sb_info *sbi)
2641 struct f2fs_super_block *sb = F2FS_RAW_SUPER(sbi);
2643 if (!(sb->feature & cpu_to_le32(F2FS_FEATURE_ENCRYPT)) &&
2644 c.feature & cpu_to_le32(F2FS_FEATURE_ENCRYPT)) {
2645 sb->feature |= cpu_to_le32(F2FS_FEATURE_ENCRYPT);
2646 MSG(0, "Info: Set Encryption feature\n");
2649 /* TODO: quota needs to allocate inode numbers */
2651 c.feature = sb->feature;
2655 update_superblock(sb, SB_MASK_ALL);
2658 int f2fs_do_mount(struct f2fs_sb_info *sbi)
2660 struct f2fs_checkpoint *cp = NULL;
2661 struct f2fs_super_block *sb = NULL;
2664 sbi->active_logs = NR_CURSEG_TYPE;
2665 ret = validate_super_block(sbi, SB0_ADDR);
2667 ret = validate_super_block(sbi, SB1_ADDR);
2671 sb = F2FS_RAW_SUPER(sbi);
2673 ret = check_sector_size(sb);
2677 print_raw_sb_info(sb);
2681 ret = get_valid_checkpoint(sbi);
2683 ERR_MSG("Can't find valid checkpoint\n");
2687 if (sanity_check_ckpt(sbi)) {
2688 ERR_MSG("Checkpoint is polluted\n");
2691 cp = F2FS_CKPT(sbi);
2693 print_ckpt_info(sbi);
2696 if (get_cp(ckpt_flags) & CP_QUOTA_NEED_FSCK_FLAG)
2702 tune_sb_features(sbi);
2704 /* precompute checksum seed for metadata */
2705 if (c.feature & cpu_to_le32(F2FS_FEATURE_INODE_CHKSUM))
2706 c.chksum_seed = f2fs_cal_crc32(~0, sb->uuid, sizeof(sb->uuid));
2708 sbi->total_valid_node_count = get_cp(valid_node_count);
2709 sbi->total_valid_inode_count = get_cp(valid_inode_count);
2710 sbi->user_block_count = get_cp(user_block_count);
2711 sbi->total_valid_block_count = get_cp(valid_block_count);
2712 sbi->last_valid_block_count = sbi->total_valid_block_count;
2713 sbi->alloc_valid_block_count = 0;
2715 if (build_segment_manager(sbi)) {
2716 ERR_MSG("build_segment_manager failed\n");
2720 if (build_node_manager(sbi)) {
2721 ERR_MSG("build_node_manager failed\n");
2725 if (!c.fix_on && (c.auto_fix || c.preen_mode)) {
2726 u32 flag = get_cp(ckpt_flags);
2728 if (flag & CP_FSCK_FLAG ||
2729 flag & CP_QUOTA_NEED_FSCK_FLAG ||
2730 (exist_qf_ino(sb) && (flag & CP_ERROR_FLAG))) {
2732 } else if (!c.preen_mode) {
2733 print_cp_state(flag);
2738 /* Check nat_bits */
2739 if (c.func == FSCK && is_set_ckpt_flags(cp, CP_NAT_BITS_FLAG)) {
2740 if (check_nat_bits(sbi, sb, cp) && c.fix_on)
2741 write_nat_bits(sbi, sb, cp, sbi->cur_cp);
2746 void f2fs_do_umount(struct f2fs_sb_info *sbi)
2748 struct sit_info *sit_i = SIT_I(sbi);
2749 struct f2fs_sm_info *sm_i = SM_I(sbi);
2750 struct f2fs_nm_info *nm_i = NM_I(sbi);
2754 if (c.func == SLOAD || c.func == FSCK)
2755 free(nm_i->nid_bitmap);
2756 free(nm_i->nat_bitmap);
2760 for (i = 0; i < TOTAL_SEGS(sbi); i++)
2761 free(sit_i->sentries[i].cur_valid_map);
2763 free(sit_i->sit_bitmap);
2764 free(sm_i->sit_info);
2767 for (i = 0; i < NR_CURSEG_TYPE; i++)
2768 free(sm_i->curseg_array[i].sum_blk);
2770 free(sm_i->curseg_array);
2774 free(sbi->raw_super);
2778 int f2fs_sparse_initialize_meta(struct f2fs_sb_info *sbi)
2780 struct f2fs_super_block *sb = sbi->raw_super;
2781 u_int32_t sit_seg_count, sit_size;
2782 u_int32_t nat_seg_count, nat_size;
2783 u_int64_t sit_seg_addr, nat_seg_addr, payload_addr;
2784 u_int32_t seg_size = 1 << get_sb(log_blocks_per_seg);
2790 sit_seg_addr = get_sb(sit_blkaddr);
2791 sit_seg_count = get_sb(segment_count_sit);
2792 sit_size = sit_seg_count * seg_size;
2794 DBG(1, "\tSparse: filling sit area at block offset: 0x%08"PRIx64" len: %u\n",
2795 sit_seg_addr, sit_size);
2796 ret = dev_fill(NULL, sit_seg_addr * F2FS_BLKSIZE,
2797 sit_size * F2FS_BLKSIZE);
2799 MSG(1, "\tError: While zeroing out the sit area "
2804 nat_seg_addr = get_sb(nat_blkaddr);
2805 nat_seg_count = get_sb(segment_count_nat);
2806 nat_size = nat_seg_count * seg_size;
2808 DBG(1, "\tSparse: filling nat area at block offset 0x%08"PRIx64" len: %u\n",
2809 nat_seg_addr, nat_size);
2810 ret = dev_fill(NULL, nat_seg_addr * F2FS_BLKSIZE,
2811 nat_size * F2FS_BLKSIZE);
2813 MSG(1, "\tError: While zeroing out the nat area "
2818 payload_addr = get_sb(segment0_blkaddr) + 1;
2820 DBG(1, "\tSparse: filling bitmap area at block offset 0x%08"PRIx64" len: %u\n",
2821 payload_addr, get_sb(cp_payload));
2822 ret = dev_fill(NULL, payload_addr * F2FS_BLKSIZE,
2823 get_sb(cp_payload) * F2FS_BLKSIZE);
2825 MSG(1, "\tError: While zeroing out the nat/sit bitmap area "
2830 payload_addr += seg_size;
2832 DBG(1, "\tSparse: filling bitmap area at block offset 0x%08"PRIx64" len: %u\n",
2833 payload_addr, get_sb(cp_payload));
2834 ret = dev_fill(NULL, payload_addr * F2FS_BLKSIZE,
2835 get_sb(cp_payload) * F2FS_BLKSIZE);
2837 MSG(1, "\tError: While zeroing out the nat/sit bitmap area "
2844 int f2fs_sparse_initialize_meta(struct f2fs_sb_info *sbi) { return 0; }