/* * QEMU Enhanced Disk Format Consistency Check * * Copyright IBM, Corp. 2010 * * Authors: * Stefan Hajnoczi * * This work is licensed under the terms of the GNU LGPL, version 2 or later. * See the COPYING.LIB file in the top-level directory. * */ #include "qed.h" typedef struct { BDRVQEDState *s; BdrvCheckResult *result; bool fix; /* whether to fix invalid offsets */ uint64_t nclusters; uint32_t *used_clusters; /* referenced cluster bitmap */ QEDRequest request; } QEDCheck; static bool qed_test_bit(uint32_t *bitmap, uint64_t n) { return !!(bitmap[n / 32] & (1 << (n % 32))); } static void qed_set_bit(uint32_t *bitmap, uint64_t n) { bitmap[n / 32] |= 1 << (n % 32); } /** * Set bitmap bits for clusters * * @check: Check structure * @offset: Starting offset in bytes * @n: Number of clusters */ static bool qed_set_used_clusters(QEDCheck *check, uint64_t offset, unsigned int n) { uint64_t cluster = qed_bytes_to_clusters(check->s, offset); unsigned int corruptions = 0; while (n-- != 0) { /* Clusters should only be referenced once */ if (qed_test_bit(check->used_clusters, cluster)) { corruptions++; } qed_set_bit(check->used_clusters, cluster); cluster++; } check->result->corruptions += corruptions; return corruptions == 0; } /** * Check an L2 table * * @ret: Number of invalid cluster offsets */ static unsigned int qed_check_l2_table(QEDCheck *check, QEDTable *table) { BDRVQEDState *s = check->s; unsigned int i, num_invalid = 0; uint64_t last_offset = 0; for (i = 0; i < s->table_nelems; i++) { uint64_t offset = table->offsets[i]; if (qed_offset_is_unalloc_cluster(offset) || qed_offset_is_zero_cluster(offset)) { continue; } check->result->bfi.allocated_clusters++; if (last_offset && (last_offset + s->header.cluster_size != offset)) { check->result->bfi.fragmented_clusters++; } last_offset = offset; /* Detect invalid cluster offset */ if (!qed_check_cluster_offset(s, offset)) { if (check->fix) { table->offsets[i] = 0; check->result->corruptions_fixed++; } else { check->result->corruptions++; } num_invalid++; continue; } qed_set_used_clusters(check, offset, 1); } return num_invalid; } /** * Descend tables and check each cluster is referenced once only */ static int qed_check_l1_table(QEDCheck *check, QEDTable *table) { BDRVQEDState *s = check->s; unsigned int i, num_invalid_l1 = 0; int ret, last_error = 0; /* Mark L1 table clusters used */ qed_set_used_clusters(check, s->header.l1_table_offset, s->header.table_size); for (i = 0; i < s->table_nelems; i++) { unsigned int num_invalid_l2; uint64_t offset = table->offsets[i]; if (qed_offset_is_unalloc_cluster(offset)) { continue; } /* Detect invalid L2 offset */ if (!qed_check_table_offset(s, offset)) { /* Clear invalid offset */ if (check->fix) { table->offsets[i] = 0; check->result->corruptions_fixed++; } else { check->result->corruptions++; } num_invalid_l1++; continue; } if (!qed_set_used_clusters(check, offset, s->header.table_size)) { continue; /* skip an invalid table */ } ret = qed_read_l2_table_sync(s, &check->request, offset); if (ret) { check->result->check_errors++; last_error = ret; continue; } num_invalid_l2 = qed_check_l2_table(check, check->request.l2_table->table); /* Write out fixed L2 table */ if (num_invalid_l2 > 0 && check->fix) { ret = qed_write_l2_table_sync(s, &check->request, 0, s->table_nelems, false); if (ret) { check->result->check_errors++; last_error = ret; continue; } } } /* Drop reference to final table */ qed_unref_l2_cache_entry(check->request.l2_table); check->request.l2_table = NULL; /* Write out fixed L1 table */ if (num_invalid_l1 > 0 && check->fix) { ret = qed_write_l1_table_sync(s, 0, s->table_nelems); if (ret) { check->result->check_errors++; last_error = ret; } } return last_error; } /** * Check for unreferenced (leaked) clusters */ static void qed_check_for_leaks(QEDCheck *check) { BDRVQEDState *s = check->s; uint64_t i; for (i = s->header.header_size; i < check->nclusters; i++) { if (!qed_test_bit(check->used_clusters, i)) { check->result->leaks++; } } } /** * Mark an image clean once it passes check or has been repaired */ static void qed_check_mark_clean(BDRVQEDState *s, BdrvCheckResult *result) { /* Skip if there were unfixable corruptions or I/O errors */ if (result->corruptions > 0 || result->check_errors > 0) { return; } /* Skip if image is already marked clean */ if (!(s->header.features & QED_F_NEED_CHECK)) { return; } /* Ensure fixes reach storage before clearing check bit */ bdrv_flush(s->bs); s->header.features &= ~QED_F_NEED_CHECK; qed_write_header_sync(s); } int qed_check(BDRVQEDState *s, BdrvCheckResult *result, bool fix) { QEDCheck check = { .s = s, .result = result, .nclusters = qed_bytes_to_clusters(s, s->file_size), .request = { .l2_table = NULL }, .fix = fix, }; int ret; check.used_clusters = g_malloc0(((check.nclusters + 31) / 32) * sizeof(check.used_clusters[0])); check.result->bfi.total_clusters = (s->header.image_size + s->header.cluster_size - 1) / s->header.cluster_size; ret = qed_check_l1_table(&check, s->l1_table); if (ret == 0) { /* Only check for leaks if entire image was scanned successfully */ qed_check_for_leaks(&check); if (fix) { qed_check_mark_clean(s, result); } } g_free(check.used_clusters); return ret; }