mirror of
https://git.hardenedbsd.org/hardenedbsd/HardenedBSD.git
synced 2024-11-21 18:50:50 +01:00
b241767f8e
The UFS1 integrity checks added in FreeBSD 14 were too aggressive for UFS1 filesystems created in FreeBSD 4 and 9 systems. This patch removes those tests which can be done safely since they are not relevant to the current implementation of UFS1. This is a follow-on report to bug report 264450 (comments 21-28). Reported by: slb@sonnet.com Tested by: slb@sonnet.com PR: 264450 MFC after: 1 week
1488 lines
37 KiB
C
1488 lines
37 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*
|
|
* Copyright (c) 1980, 1986, 1993
|
|
* The Regents of the University of California. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/time.h>
|
|
#include <sys/types.h>
|
|
#include <sys/sysctl.h>
|
|
#include <sys/disk.h>
|
|
#include <sys/disklabel.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <ufs/ufs/dinode.h>
|
|
#include <ufs/ufs/dir.h>
|
|
#include <ufs/ffs/fs.h>
|
|
|
|
#include <err.h>
|
|
#include <errno.h>
|
|
#include <string.h>
|
|
#include <ctype.h>
|
|
#include <fstab.h>
|
|
#include <stdint.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <time.h>
|
|
#include <unistd.h>
|
|
|
|
#include "fsck.h"
|
|
|
|
int sujrecovery = 0;
|
|
|
|
static struct bufarea *allocbuf(const char *);
|
|
static void cg_write(struct bufarea *);
|
|
static void slowio_start(void);
|
|
static void slowio_end(void);
|
|
static void printIOstats(void);
|
|
|
|
static long diskreads, totaldiskreads, totalreads; /* Disk cache statistics */
|
|
static struct timespec startpass, finishpass;
|
|
struct timeval slowio_starttime;
|
|
int slowio_delay_usec = 10000; /* Initial IO delay for background fsck */
|
|
int slowio_pollcnt;
|
|
static struct bufarea cgblk; /* backup buffer for cylinder group blocks */
|
|
static struct bufarea failedbuf; /* returned by failed getdatablk() */
|
|
static TAILQ_HEAD(bufqueue, bufarea) bufqueuehd; /* head of buffer cache LRU */
|
|
static LIST_HEAD(bufhash, bufarea) bufhashhd[HASHSIZE]; /* buffer hash list */
|
|
static struct bufhash freebufs; /* unused buffers */
|
|
static int numbufs; /* size of buffer cache */
|
|
static int cachelookups; /* number of cache lookups */
|
|
static int cachereads; /* number of cache reads */
|
|
static int flushtries; /* number of tries to reclaim memory */
|
|
|
|
char *buftype[BT_NUMBUFTYPES] = BT_NAMES;
|
|
|
|
void
|
|
fsutilinit(void)
|
|
{
|
|
diskreads = totaldiskreads = totalreads = 0;
|
|
bzero(&startpass, sizeof(struct timespec));
|
|
bzero(&finishpass, sizeof(struct timespec));
|
|
bzero(&slowio_starttime, sizeof(struct timeval));
|
|
slowio_delay_usec = 10000;
|
|
slowio_pollcnt = 0;
|
|
flushtries = 0;
|
|
}
|
|
|
|
int
|
|
ftypeok(union dinode *dp)
|
|
{
|
|
switch (DIP(dp, di_mode) & IFMT) {
|
|
|
|
case IFDIR:
|
|
case IFREG:
|
|
case IFBLK:
|
|
case IFCHR:
|
|
case IFLNK:
|
|
case IFSOCK:
|
|
case IFIFO:
|
|
return (1);
|
|
|
|
default:
|
|
if (debug)
|
|
printf("bad file type 0%o\n", DIP(dp, di_mode));
|
|
return (0);
|
|
}
|
|
}
|
|
|
|
int
|
|
reply(const char *question)
|
|
{
|
|
int persevere;
|
|
char c;
|
|
|
|
if (preen)
|
|
pfatal("INTERNAL ERROR: GOT TO reply()");
|
|
persevere = strcmp(question, "CONTINUE") == 0 ||
|
|
strcmp(question, "LOOK FOR ALTERNATE SUPERBLOCKS") == 0;
|
|
printf("\n");
|
|
if (!persevere && (nflag || (fswritefd < 0 && bkgrdflag == 0))) {
|
|
printf("%s? no\n\n", question);
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
if (yflag || (persevere && nflag)) {
|
|
printf("%s? yes\n\n", question);
|
|
return (1);
|
|
}
|
|
do {
|
|
printf("%s? [yn] ", question);
|
|
(void) fflush(stdout);
|
|
c = getc(stdin);
|
|
while (c != '\n' && getc(stdin) != '\n') {
|
|
if (feof(stdin)) {
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
}
|
|
} while (c != 'y' && c != 'Y' && c != 'n' && c != 'N');
|
|
printf("\n");
|
|
if (c == 'y' || c == 'Y')
|
|
return (1);
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Look up state information for an inode.
|
|
*/
|
|
struct inostat *
|
|
inoinfo(ino_t inum)
|
|
{
|
|
static struct inostat unallocated = { USTATE, 0, 0, 0 };
|
|
struct inostatlist *ilp;
|
|
int iloff;
|
|
|
|
if (inum >= maxino)
|
|
errx(EEXIT, "inoinfo: inumber %ju out of range",
|
|
(uintmax_t)inum);
|
|
ilp = &inostathead[inum / sblock.fs_ipg];
|
|
iloff = inum % sblock.fs_ipg;
|
|
if (iloff >= ilp->il_numalloced)
|
|
return (&unallocated);
|
|
return (&ilp->il_stat[iloff]);
|
|
}
|
|
|
|
/*
|
|
* Malloc buffers and set up cache.
|
|
*/
|
|
void
|
|
bufinit(void)
|
|
{
|
|
int i;
|
|
|
|
initbarea(&failedbuf, BT_UNKNOWN);
|
|
failedbuf.b_errs = -1;
|
|
failedbuf.b_un.b_buf = NULL;
|
|
if ((cgblk.b_un.b_buf = Balloc((unsigned int)sblock.fs_bsize)) == NULL)
|
|
errx(EEXIT, "Initial malloc(%d) failed", sblock.fs_bsize);
|
|
initbarea(&cgblk, BT_CYLGRP);
|
|
numbufs = cachelookups = cachereads = 0;
|
|
TAILQ_INIT(&bufqueuehd);
|
|
LIST_INIT(&freebufs);
|
|
for (i = 0; i < HASHSIZE; i++)
|
|
LIST_INIT(&bufhashhd[i]);
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
readtime[i].tv_sec = totalreadtime[i].tv_sec = 0;
|
|
readtime[i].tv_nsec = totalreadtime[i].tv_nsec = 0;
|
|
readcnt[i] = totalreadcnt[i] = 0;
|
|
}
|
|
}
|
|
|
|
static struct bufarea *
|
|
allocbuf(const char *failreason)
|
|
{
|
|
struct bufarea *bp;
|
|
char *bufp;
|
|
|
|
bp = (struct bufarea *)Malloc(sizeof(struct bufarea));
|
|
bufp = Balloc((unsigned int)sblock.fs_bsize);
|
|
if (bp == NULL || bufp == NULL) {
|
|
errx(EEXIT, "%s", failreason);
|
|
/* NOTREACHED */
|
|
}
|
|
numbufs++;
|
|
bp->b_un.b_buf = bufp;
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
initbarea(bp, BT_UNKNOWN);
|
|
return (bp);
|
|
}
|
|
|
|
/*
|
|
* Manage cylinder group buffers.
|
|
*
|
|
* Use getblk() here rather than cgget() because the cylinder group
|
|
* may be corrupted but we want it anyway so we can fix it.
|
|
*/
|
|
static struct bufarea *cgbufs; /* header for cylinder group cache */
|
|
static int flushtries; /* number of tries to reclaim memory */
|
|
|
|
struct bufarea *
|
|
cglookup(int cg)
|
|
{
|
|
struct bufarea *cgbp;
|
|
struct cg *cgp;
|
|
|
|
if ((unsigned) cg >= sblock.fs_ncg)
|
|
errx(EEXIT, "cglookup: out of range cylinder group %d", cg);
|
|
if (cgbufs == NULL) {
|
|
cgbufs = Calloc(sblock.fs_ncg, sizeof(struct bufarea));
|
|
if (cgbufs == NULL)
|
|
errx(EEXIT, "Cannot allocate cylinder group buffers");
|
|
}
|
|
cgbp = &cgbufs[cg];
|
|
if (cgbp->b_un.b_cg != NULL)
|
|
return (cgbp);
|
|
cgp = NULL;
|
|
if (flushtries == 0)
|
|
cgp = Balloc((unsigned int)sblock.fs_cgsize);
|
|
if (cgp == NULL) {
|
|
if (sujrecovery)
|
|
errx(EEXIT,"Ran out of memory during journal recovery");
|
|
flush(fswritefd, &cgblk);
|
|
getblk(&cgblk, cgtod(&sblock, cg), sblock.fs_cgsize);
|
|
return (&cgblk);
|
|
}
|
|
cgbp->b_un.b_cg = cgp;
|
|
initbarea(cgbp, BT_CYLGRP);
|
|
getblk(cgbp, cgtod(&sblock, cg), sblock.fs_cgsize);
|
|
return (cgbp);
|
|
}
|
|
|
|
/*
|
|
* Mark a cylinder group buffer as dirty.
|
|
* Update its check-hash if they are enabled.
|
|
*/
|
|
void
|
|
cgdirty(struct bufarea *cgbp)
|
|
{
|
|
struct cg *cg;
|
|
|
|
cg = cgbp->b_un.b_cg;
|
|
if ((sblock.fs_metackhash & CK_CYLGRP) != 0) {
|
|
cg->cg_ckhash = 0;
|
|
cg->cg_ckhash =
|
|
calculate_crc32c(~0L, (void *)cg, sblock.fs_cgsize);
|
|
}
|
|
dirty(cgbp);
|
|
}
|
|
|
|
/*
|
|
* Attempt to flush a cylinder group cache entry.
|
|
* Return whether the flush was successful.
|
|
*/
|
|
int
|
|
flushentry(void)
|
|
{
|
|
struct bufarea *cgbp;
|
|
|
|
if (sujrecovery || flushtries == sblock.fs_ncg || cgbufs == NULL)
|
|
return (0);
|
|
cgbp = &cgbufs[flushtries++];
|
|
if (cgbp->b_un.b_cg == NULL)
|
|
return (0);
|
|
flush(fswritefd, cgbp);
|
|
free(cgbp->b_un.b_buf);
|
|
cgbp->b_un.b_buf = NULL;
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* Manage a cache of filesystem disk blocks.
|
|
*/
|
|
struct bufarea *
|
|
getdatablk(ufs2_daddr_t blkno, long size, int type)
|
|
{
|
|
struct bufarea *bp;
|
|
struct bufhash *bhdp;
|
|
|
|
cachelookups++;
|
|
/*
|
|
* If out of range, return empty buffer with b_err == -1
|
|
*
|
|
* Skip check for inodes because chkrange() considers
|
|
* metadata areas invalid to write data.
|
|
*/
|
|
if (type != BT_INODES && chkrange(blkno, size / sblock.fs_fsize)) {
|
|
failedbuf.b_refcnt++;
|
|
return (&failedbuf);
|
|
}
|
|
bhdp = &bufhashhd[HASH(blkno)];
|
|
LIST_FOREACH(bp, bhdp, b_hash)
|
|
if (bp->b_bno == fsbtodb(&sblock, blkno)) {
|
|
if (debug && bp->b_size != size) {
|
|
prtbuf(bp, "getdatablk: size mismatch");
|
|
pfatal("getdatablk: b_size %d != size %ld\n",
|
|
bp->b_size, size);
|
|
}
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
goto foundit;
|
|
}
|
|
/*
|
|
* Move long-term busy buffer back to the front of the LRU so we
|
|
* do not endless inspect them for recycling.
|
|
*/
|
|
bp = TAILQ_LAST(&bufqueuehd, bufqueue);
|
|
if (bp != NULL && bp->b_refcnt != 0) {
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
}
|
|
/*
|
|
* Allocate up to the minimum number of buffers before
|
|
* considering recycling any of them.
|
|
*/
|
|
if (size > sblock.fs_bsize)
|
|
errx(EEXIT, "Excessive buffer size %ld > %d\n", size,
|
|
sblock.fs_bsize);
|
|
if ((bp = LIST_FIRST(&freebufs)) != NULL) {
|
|
LIST_REMOVE(bp, b_hash);
|
|
} else if (numbufs < MINBUFS) {
|
|
bp = allocbuf("cannot create minimal buffer pool");
|
|
} else if (sujrecovery) {
|
|
/*
|
|
* SUJ recovery does not want anything written until it
|
|
* has successfully completed (so it can fail back to
|
|
* full fsck). Thus, we can only recycle clean buffers.
|
|
*/
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if ((bp->b_flags & B_DIRTY) == 0 && bp->b_refcnt == 0)
|
|
break;
|
|
if (bp == NULL)
|
|
bp = allocbuf("Ran out of memory during "
|
|
"journal recovery");
|
|
else
|
|
LIST_REMOVE(bp, b_hash);
|
|
} else {
|
|
/*
|
|
* Recycle oldest non-busy buffer.
|
|
*/
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if (bp->b_refcnt == 0)
|
|
break;
|
|
if (bp == NULL)
|
|
bp = allocbuf("Ran out of memory for buffers");
|
|
else
|
|
LIST_REMOVE(bp, b_hash);
|
|
}
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
flush(fswritefd, bp);
|
|
bp->b_type = type;
|
|
LIST_INSERT_HEAD(bhdp, bp, b_hash);
|
|
getblk(bp, blkno, size);
|
|
cachereads++;
|
|
/* fall through */
|
|
foundit:
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
if (debug && bp->b_type != type) {
|
|
printf("getdatablk: buffer type changed to %s",
|
|
BT_BUFTYPE(type));
|
|
prtbuf(bp, "");
|
|
}
|
|
if (bp->b_errs == 0)
|
|
bp->b_refcnt++;
|
|
return (bp);
|
|
}
|
|
|
|
void
|
|
getblk(struct bufarea *bp, ufs2_daddr_t blk, long size)
|
|
{
|
|
ufs2_daddr_t dblk;
|
|
struct timespec start, finish;
|
|
|
|
dblk = fsbtodb(&sblock, blk);
|
|
if (bp->b_bno == dblk) {
|
|
totalreads++;
|
|
} else {
|
|
if (debug) {
|
|
readcnt[bp->b_type]++;
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &start);
|
|
}
|
|
bp->b_errs = blread(fsreadfd, bp->b_un.b_buf, dblk, size);
|
|
if (debug) {
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &finish);
|
|
timespecsub(&finish, &start, &finish);
|
|
timespecadd(&readtime[bp->b_type], &finish,
|
|
&readtime[bp->b_type]);
|
|
}
|
|
bp->b_bno = dblk;
|
|
bp->b_size = size;
|
|
}
|
|
}
|
|
|
|
void
|
|
brelse(struct bufarea *bp)
|
|
{
|
|
|
|
if (bp->b_refcnt <= 0)
|
|
prtbuf(bp, "brelse: buffer with negative reference count");
|
|
bp->b_refcnt--;
|
|
}
|
|
|
|
void
|
|
binval(struct bufarea *bp)
|
|
{
|
|
|
|
bp->b_flags &= ~B_DIRTY;
|
|
LIST_REMOVE(bp, b_hash);
|
|
LIST_INSERT_HEAD(&freebufs, bp, b_hash);
|
|
}
|
|
|
|
void
|
|
flush(int fd, struct bufarea *bp)
|
|
{
|
|
struct inode ip;
|
|
|
|
if ((bp->b_flags & B_DIRTY) == 0)
|
|
return;
|
|
bp->b_flags &= ~B_DIRTY;
|
|
if (fswritefd < 0) {
|
|
pfatal("WRITING IN READ_ONLY MODE.\n");
|
|
return;
|
|
}
|
|
if (bp->b_errs != 0)
|
|
pfatal("WRITING %sZERO'ED BLOCK %lld TO DISK\n",
|
|
(bp->b_errs == bp->b_size / dev_bsize) ? "" : "PARTIALLY ",
|
|
(long long)bp->b_bno);
|
|
bp->b_errs = 0;
|
|
/*
|
|
* Write using the appropriate function.
|
|
*/
|
|
switch (bp->b_type) {
|
|
case BT_SUPERBLK:
|
|
if (bp != &sblk)
|
|
pfatal("BUFFER %p DOES NOT MATCH SBLK %p\n",
|
|
bp, &sblk);
|
|
/*
|
|
* Superblocks are always pre-copied so we do not need
|
|
* to check them for copy-on-write.
|
|
*/
|
|
if (sbput(fd, bp->b_un.b_fs, 0) == 0)
|
|
fsmodified = 1;
|
|
break;
|
|
case BT_CYLGRP:
|
|
/*
|
|
* Cylinder groups are always pre-copied so we do not
|
|
* need to check them for copy-on-write.
|
|
*/
|
|
if (sujrecovery)
|
|
cg_write(bp);
|
|
if (cgput(fswritefd, &sblock, bp->b_un.b_cg) == 0)
|
|
fsmodified = 1;
|
|
break;
|
|
case BT_INODES:
|
|
if (debug && sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
struct ufs2_dinode *dp = bp->b_un.b_dinode2;
|
|
int i;
|
|
|
|
for (i = 0; i < bp->b_size; dp++, i += sizeof(*dp)) {
|
|
if (ffs_verify_dinode_ckhash(&sblock, dp) == 0)
|
|
continue;
|
|
pwarn("flush: INODE CHECK-HASH FAILED");
|
|
ip.i_bp = bp;
|
|
ip.i_dp = (union dinode *)dp;
|
|
ip.i_number = bp->b_index + (i / sizeof(*dp));
|
|
prtinode(&ip);
|
|
if (preen || reply("FIX") != 0) {
|
|
if (preen)
|
|
printf(" (FIXED)\n");
|
|
ffs_update_dinode_ckhash(&sblock, dp);
|
|
inodirty(&ip);
|
|
}
|
|
}
|
|
}
|
|
/* FALLTHROUGH */
|
|
default:
|
|
copyonwrite(&sblock, bp, std_checkblkavail);
|
|
blwrite(fd, bp->b_un.b_buf, bp->b_bno, bp->b_size);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If there are any snapshots, ensure that all the blocks that they
|
|
* care about have been copied, then release the snapshot inodes.
|
|
* These operations need to be done before we rebuild the cylinder
|
|
* groups so that any block allocations are properly recorded.
|
|
* Since all the cylinder group maps have already been copied in
|
|
* the snapshots, no further snapshot copies will need to be done.
|
|
*/
|
|
void
|
|
snapflush(ufs2_daddr_t (*checkblkavail)(ufs2_daddr_t, long))
|
|
{
|
|
struct bufarea *bp;
|
|
int cnt;
|
|
|
|
if (snapcnt > 0) {
|
|
if (debug)
|
|
printf("Check for snapshot copies\n");
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if ((bp->b_flags & B_DIRTY) != 0)
|
|
copyonwrite(&sblock, bp, checkblkavail);
|
|
for (cnt = 0; cnt < snapcnt; cnt++)
|
|
irelse(&snaplist[cnt]);
|
|
snapcnt = 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Journaled soft updates does not maintain cylinder group summary
|
|
* information during cleanup, so this routine recalculates the summary
|
|
* information and updates the superblock summary in preparation for
|
|
* writing out the cylinder group.
|
|
*/
|
|
static void
|
|
cg_write(struct bufarea *bp)
|
|
{
|
|
ufs1_daddr_t fragno, cgbno, maxbno;
|
|
u_int8_t *blksfree;
|
|
struct csum *csp;
|
|
struct cg *cgp;
|
|
int blk;
|
|
int i;
|
|
|
|
/*
|
|
* Fix the frag and cluster summary.
|
|
*/
|
|
cgp = bp->b_un.b_cg;
|
|
cgp->cg_cs.cs_nbfree = 0;
|
|
cgp->cg_cs.cs_nffree = 0;
|
|
bzero(&cgp->cg_frsum, sizeof(cgp->cg_frsum));
|
|
maxbno = fragstoblks(&sblock, sblock.fs_fpg);
|
|
if (sblock.fs_contigsumsize > 0) {
|
|
for (i = 1; i <= sblock.fs_contigsumsize; i++)
|
|
cg_clustersum(cgp)[i] = 0;
|
|
bzero(cg_clustersfree(cgp), howmany(maxbno, CHAR_BIT));
|
|
}
|
|
blksfree = cg_blksfree(cgp);
|
|
for (cgbno = 0; cgbno < maxbno; cgbno++) {
|
|
if (ffs_isfreeblock(&sblock, blksfree, cgbno))
|
|
continue;
|
|
if (ffs_isblock(&sblock, blksfree, cgbno)) {
|
|
ffs_clusteracct(&sblock, cgp, cgbno, 1);
|
|
cgp->cg_cs.cs_nbfree++;
|
|
continue;
|
|
}
|
|
fragno = blkstofrags(&sblock, cgbno);
|
|
blk = blkmap(&sblock, blksfree, fragno);
|
|
ffs_fragacct(&sblock, blk, cgp->cg_frsum, 1);
|
|
for (i = 0; i < sblock.fs_frag; i++)
|
|
if (isset(blksfree, fragno + i))
|
|
cgp->cg_cs.cs_nffree++;
|
|
}
|
|
/*
|
|
* Update the superblock cg summary from our now correct values
|
|
* before writing the block.
|
|
*/
|
|
csp = &sblock.fs_cs(&sblock, cgp->cg_cgx);
|
|
sblock.fs_cstotal.cs_ndir += cgp->cg_cs.cs_ndir - csp->cs_ndir;
|
|
sblock.fs_cstotal.cs_nbfree += cgp->cg_cs.cs_nbfree - csp->cs_nbfree;
|
|
sblock.fs_cstotal.cs_nifree += cgp->cg_cs.cs_nifree - csp->cs_nifree;
|
|
sblock.fs_cstotal.cs_nffree += cgp->cg_cs.cs_nffree - csp->cs_nffree;
|
|
sblock.fs_cs(&sblock, cgp->cg_cgx) = cgp->cg_cs;
|
|
}
|
|
|
|
void
|
|
rwerror(const char *mesg, ufs2_daddr_t blk)
|
|
{
|
|
|
|
if (bkgrdcheck)
|
|
exit(EEXIT);
|
|
if (preen == 0)
|
|
printf("\n");
|
|
pfatal("CANNOT %s: %ld", mesg, (long)blk);
|
|
if (reply("CONTINUE") == 0)
|
|
exit(EEXIT);
|
|
}
|
|
|
|
void
|
|
ckfini(int markclean)
|
|
{
|
|
struct bufarea *bp, *nbp;
|
|
int ofsmodified, cnt, cg;
|
|
|
|
if (bkgrdflag) {
|
|
if ((!(sblock.fs_flags & FS_UNCLEAN)) != markclean) {
|
|
cmd.value = FS_UNCLEAN;
|
|
cmd.size = markclean ? -1 : 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FILE SYSTEM DIRTY FLAG\n");
|
|
if (!preen) {
|
|
printf("\n***** FILE SYSTEM MARKED %s *****\n",
|
|
markclean ? "CLEAN" : "DIRTY");
|
|
if (!markclean)
|
|
rerun = 1;
|
|
}
|
|
} else if (!preen && !markclean) {
|
|
printf("\n***** FILE SYSTEM STILL DIRTY *****\n");
|
|
rerun = 1;
|
|
}
|
|
bkgrdflag = 0;
|
|
}
|
|
if (debug && cachelookups > 0)
|
|
printf("cache with %d buffers missed %d of %d (%d%%)\n",
|
|
numbufs, cachereads, cachelookups,
|
|
(int)(cachereads * 100 / cachelookups));
|
|
if (fswritefd < 0) {
|
|
(void)close(fsreadfd);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* To remain idempotent with partial truncations the buffers
|
|
* must be flushed in this order:
|
|
* 1) cylinder groups (bitmaps)
|
|
* 2) indirect, directory, external attribute, and data blocks
|
|
* 3) inode blocks
|
|
* 4) superblock
|
|
* This ordering preserves access to the modified pointers
|
|
* until they are freed.
|
|
*/
|
|
/* Step 1: cylinder groups */
|
|
if (debug)
|
|
printf("Flush Cylinder groups\n");
|
|
if (cgbufs != NULL) {
|
|
for (cnt = 0; cnt < sblock.fs_ncg; cnt++) {
|
|
if (cgbufs[cnt].b_un.b_cg == NULL)
|
|
continue;
|
|
flush(fswritefd, &cgbufs[cnt]);
|
|
free(cgbufs[cnt].b_un.b_cg);
|
|
}
|
|
free(cgbufs);
|
|
cgbufs = NULL;
|
|
}
|
|
flush(fswritefd, &cgblk);
|
|
free(cgblk.b_un.b_buf);
|
|
cgblk.b_un.b_buf = NULL;
|
|
cnt = 0;
|
|
/* Step 2: indirect, directory, external attribute, and data blocks */
|
|
if (debug)
|
|
printf("Flush indirect, directory, external attribute, "
|
|
"and data blocks\n");
|
|
if (pdirbp != NULL) {
|
|
brelse(pdirbp);
|
|
pdirbp = NULL;
|
|
}
|
|
TAILQ_FOREACH_REVERSE_SAFE(bp, &bufqueuehd, bufqueue, b_list, nbp) {
|
|
switch (bp->b_type) {
|
|
/* These should not be in the buffer cache list */
|
|
case BT_UNKNOWN:
|
|
case BT_SUPERBLK:
|
|
case BT_CYLGRP:
|
|
default:
|
|
prtbuf(bp,"ckfini: improper buffer type on cache list");
|
|
continue;
|
|
/* These are the ones to flush in this step */
|
|
case BT_LEVEL1:
|
|
case BT_LEVEL2:
|
|
case BT_LEVEL3:
|
|
case BT_EXTATTR:
|
|
case BT_DIRDATA:
|
|
case BT_DATA:
|
|
break;
|
|
/* These are the ones to flush in the next step */
|
|
case BT_INODES:
|
|
continue;
|
|
}
|
|
if (debug && bp->b_refcnt != 0)
|
|
prtbuf(bp, "ckfini: clearing in-use buffer");
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
LIST_REMOVE(bp, b_hash);
|
|
cnt++;
|
|
flush(fswritefd, bp);
|
|
free(bp->b_un.b_buf);
|
|
free((char *)bp);
|
|
}
|
|
/* Step 3: inode blocks */
|
|
if (debug)
|
|
printf("Flush inode blocks\n");
|
|
if (icachebp != NULL) {
|
|
brelse(icachebp);
|
|
icachebp = NULL;
|
|
}
|
|
TAILQ_FOREACH_REVERSE_SAFE(bp, &bufqueuehd, bufqueue, b_list, nbp) {
|
|
if (debug && bp->b_refcnt != 0)
|
|
prtbuf(bp, "ckfini: clearing in-use buffer");
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
LIST_REMOVE(bp, b_hash);
|
|
cnt++;
|
|
flush(fswritefd, bp);
|
|
free(bp->b_un.b_buf);
|
|
free((char *)bp);
|
|
}
|
|
if (numbufs != cnt)
|
|
errx(EEXIT, "panic: lost %d buffers", numbufs - cnt);
|
|
/* Step 4: superblock */
|
|
if (debug)
|
|
printf("Flush the superblock\n");
|
|
flush(fswritefd, &sblk);
|
|
if (havesb && cursnapshot == 0 &&
|
|
sblk.b_bno != sblock.fs_sblockloc / dev_bsize) {
|
|
if (preen || reply("UPDATE STANDARD SUPERBLOCK")) {
|
|
/* Change write destination to standard superblock */
|
|
sblock.fs_sblockactualloc = sblock.fs_sblockloc;
|
|
sblk.b_bno = sblock.fs_sblockloc / dev_bsize;
|
|
sbdirty();
|
|
flush(fswritefd, &sblk);
|
|
} else {
|
|
markclean = 0;
|
|
}
|
|
}
|
|
if (cursnapshot == 0 && sblock.fs_clean != markclean) {
|
|
if ((sblock.fs_clean = markclean) != 0) {
|
|
sblock.fs_flags &= ~(FS_UNCLEAN | FS_NEEDSFSCK);
|
|
sblock.fs_pendingblocks = 0;
|
|
sblock.fs_pendinginodes = 0;
|
|
}
|
|
sbdirty();
|
|
ofsmodified = fsmodified;
|
|
flush(fswritefd, &sblk);
|
|
fsmodified = ofsmodified;
|
|
if (!preen) {
|
|
printf("\n***** FILE SYSTEM MARKED %s *****\n",
|
|
markclean ? "CLEAN" : "DIRTY");
|
|
if (!markclean)
|
|
rerun = 1;
|
|
}
|
|
} else if (!preen) {
|
|
if (markclean) {
|
|
printf("\n***** FILE SYSTEM IS CLEAN *****\n");
|
|
} else {
|
|
printf("\n***** FILE SYSTEM STILL DIRTY *****\n");
|
|
rerun = 1;
|
|
}
|
|
}
|
|
/*
|
|
* Free allocated tracking structures.
|
|
*/
|
|
if (blockmap != NULL)
|
|
free(blockmap);
|
|
blockmap = NULL;
|
|
if (inostathead != NULL) {
|
|
for (cg = 0; cg < sblock.fs_ncg; cg++)
|
|
if (inostathead[cg].il_stat != NULL)
|
|
free((char *)inostathead[cg].il_stat);
|
|
free(inostathead);
|
|
}
|
|
inostathead = NULL;
|
|
inocleanup();
|
|
finalIOstats();
|
|
(void)close(fsreadfd);
|
|
(void)close(fswritefd);
|
|
}
|
|
|
|
/*
|
|
* Print out I/O statistics.
|
|
*/
|
|
void
|
|
IOstats(char *what)
|
|
{
|
|
int i;
|
|
|
|
if (debug == 0)
|
|
return;
|
|
if (diskreads == 0) {
|
|
printf("%s: no I/O\n\n", what);
|
|
return;
|
|
}
|
|
if (startpass.tv_sec == 0)
|
|
startpass = startprog;
|
|
printf("%s: I/O statistics\n", what);
|
|
printIOstats();
|
|
totaldiskreads += diskreads;
|
|
diskreads = 0;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
timespecadd(&totalreadtime[i], &readtime[i], &totalreadtime[i]);
|
|
totalreadcnt[i] += readcnt[i];
|
|
readtime[i].tv_sec = readtime[i].tv_nsec = 0;
|
|
readcnt[i] = 0;
|
|
}
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &startpass);
|
|
}
|
|
|
|
void
|
|
finalIOstats(void)
|
|
{
|
|
int i;
|
|
|
|
if (debug == 0)
|
|
return;
|
|
printf("Final I/O statistics\n");
|
|
totaldiskreads += diskreads;
|
|
diskreads = totaldiskreads;
|
|
startpass = startprog;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
timespecadd(&totalreadtime[i], &readtime[i], &totalreadtime[i]);
|
|
totalreadcnt[i] += readcnt[i];
|
|
readtime[i] = totalreadtime[i];
|
|
readcnt[i] = totalreadcnt[i];
|
|
}
|
|
printIOstats();
|
|
}
|
|
|
|
static void printIOstats(void)
|
|
{
|
|
long long msec, totalmsec;
|
|
int i;
|
|
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &finishpass);
|
|
timespecsub(&finishpass, &startpass, &finishpass);
|
|
printf("Running time: %jd.%03ld sec\n",
|
|
(intmax_t)finishpass.tv_sec, finishpass.tv_nsec / 1000000);
|
|
printf("buffer reads by type:\n");
|
|
for (totalmsec = 0, i = 0; i < BT_NUMBUFTYPES; i++)
|
|
totalmsec += readtime[i].tv_sec * 1000 +
|
|
readtime[i].tv_nsec / 1000000;
|
|
if (totalmsec == 0)
|
|
totalmsec = 1;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
if (readcnt[i] == 0)
|
|
continue;
|
|
msec =
|
|
readtime[i].tv_sec * 1000 + readtime[i].tv_nsec / 1000000;
|
|
printf("%21s:%8ld %2ld.%ld%% %4jd.%03ld sec %2lld.%lld%%\n",
|
|
buftype[i], readcnt[i], readcnt[i] * 100 / diskreads,
|
|
(readcnt[i] * 1000 / diskreads) % 10,
|
|
(intmax_t)readtime[i].tv_sec, readtime[i].tv_nsec / 1000000,
|
|
msec * 100 / totalmsec, (msec * 1000 / totalmsec) % 10);
|
|
}
|
|
printf("\n");
|
|
}
|
|
|
|
int
|
|
blread(int fd, char *buf, ufs2_daddr_t blk, long size)
|
|
{
|
|
char *cp;
|
|
int i, errs;
|
|
off_t offset;
|
|
|
|
offset = blk;
|
|
offset *= dev_bsize;
|
|
if (bkgrdflag)
|
|
slowio_start();
|
|
totalreads++;
|
|
diskreads++;
|
|
if (pread(fd, buf, (int)size, offset) == size) {
|
|
if (bkgrdflag)
|
|
slowio_end();
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* This is handled specially here instead of in rwerror because
|
|
* rwerror is used for all sorts of errors, not just true read/write
|
|
* errors. It should be refactored and fixed.
|
|
*/
|
|
if (surrender) {
|
|
pfatal("CANNOT READ_BLK: %ld", (long)blk);
|
|
errx(EEXIT, "ABORTING DUE TO READ ERRORS");
|
|
} else
|
|
rwerror("READ BLK", blk);
|
|
|
|
errs = 0;
|
|
memset(buf, 0, (size_t)size);
|
|
printf("THE FOLLOWING DISK SECTORS COULD NOT BE READ:");
|
|
for (cp = buf, i = 0; i < size; i += secsize, cp += secsize) {
|
|
if (pread(fd, cp, (int)secsize, offset + i) != secsize) {
|
|
if (secsize != dev_bsize && dev_bsize != 1)
|
|
printf(" %jd (%jd),",
|
|
(intmax_t)(blk * dev_bsize + i) / secsize,
|
|
(intmax_t)blk + i / dev_bsize);
|
|
else
|
|
printf(" %jd,", (intmax_t)blk + i / dev_bsize);
|
|
errs++;
|
|
}
|
|
}
|
|
printf("\n");
|
|
if (errs)
|
|
resolved = 0;
|
|
return (errs);
|
|
}
|
|
|
|
void
|
|
blwrite(int fd, char *buf, ufs2_daddr_t blk, ssize_t size)
|
|
{
|
|
int i;
|
|
char *cp;
|
|
off_t offset;
|
|
|
|
if (fd < 0)
|
|
return;
|
|
offset = blk;
|
|
offset *= dev_bsize;
|
|
if (pwrite(fd, buf, size, offset) == size) {
|
|
fsmodified = 1;
|
|
return;
|
|
}
|
|
resolved = 0;
|
|
rwerror("WRITE BLK", blk);
|
|
printf("THE FOLLOWING SECTORS COULD NOT BE WRITTEN:");
|
|
for (cp = buf, i = 0; i < size; i += dev_bsize, cp += dev_bsize)
|
|
if (pwrite(fd, cp, dev_bsize, offset + i) != dev_bsize)
|
|
printf(" %jd,", (intmax_t)blk + i / dev_bsize);
|
|
printf("\n");
|
|
return;
|
|
}
|
|
|
|
void
|
|
blerase(int fd, ufs2_daddr_t blk, long size)
|
|
{
|
|
off_t ioarg[2];
|
|
|
|
if (fd < 0)
|
|
return;
|
|
ioarg[0] = blk * dev_bsize;
|
|
ioarg[1] = size;
|
|
ioctl(fd, DIOCGDELETE, ioarg);
|
|
/* we don't really care if we succeed or not */
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Fill a contiguous region with all-zeroes. Note ZEROBUFSIZE is by
|
|
* definition a multiple of dev_bsize.
|
|
*/
|
|
void
|
|
blzero(int fd, ufs2_daddr_t blk, long size)
|
|
{
|
|
static char *zero;
|
|
off_t offset, len;
|
|
|
|
if (fd < 0)
|
|
return;
|
|
if (zero == NULL) {
|
|
zero = Balloc(ZEROBUFSIZE);
|
|
if (zero == NULL)
|
|
errx(EEXIT, "cannot allocate buffer pool");
|
|
}
|
|
offset = blk * dev_bsize;
|
|
if (lseek(fd, offset, 0) < 0)
|
|
rwerror("SEEK BLK", blk);
|
|
while (size > 0) {
|
|
len = MIN(ZEROBUFSIZE, size);
|
|
if (write(fd, zero, len) != len)
|
|
rwerror("WRITE BLK", blk);
|
|
blk += len / dev_bsize;
|
|
size -= len;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Verify cylinder group's magic number and other parameters. If the
|
|
* test fails, offer an option to rebuild the whole cylinder group.
|
|
*
|
|
* Return 1 if the cylinder group is good or return 0 if it is bad.
|
|
*/
|
|
#undef CHK
|
|
#define CHK(lhs, op, rhs, fmt) \
|
|
if (lhs op rhs) { \
|
|
pwarn("UFS%d cylinder group %d failed: " \
|
|
"%s (" #fmt ") %s %s (" #fmt ")\n", \
|
|
sblock.fs_magic == FS_UFS1_MAGIC ? 1 : 2, cg, \
|
|
#lhs, (intmax_t)lhs, #op, #rhs, (intmax_t)rhs); \
|
|
error = 1; \
|
|
}
|
|
int
|
|
check_cgmagic(int cg, struct bufarea *cgbp)
|
|
{
|
|
struct cg *cgp = cgbp->b_un.b_cg;
|
|
uint32_t cghash, calchash;
|
|
static int prevfailcg = -1;
|
|
long start;
|
|
int error;
|
|
|
|
/*
|
|
* Extended cylinder group checks.
|
|
*/
|
|
calchash = cgp->cg_ckhash;
|
|
if ((sblock.fs_metackhash & CK_CYLGRP) != 0 &&
|
|
(ckhashadd & CK_CYLGRP) == 0) {
|
|
cghash = cgp->cg_ckhash;
|
|
cgp->cg_ckhash = 0;
|
|
calchash = calculate_crc32c(~0L, (void *)cgp, sblock.fs_cgsize);
|
|
cgp->cg_ckhash = cghash;
|
|
}
|
|
error = 0;
|
|
CHK(cgp->cg_ckhash, !=, calchash, "%jd");
|
|
CHK(cg_chkmagic(cgp), ==, 0, "%jd");
|
|
CHK(cgp->cg_cgx, !=, cg, "%jd");
|
|
CHK(cgp->cg_ndblk, >, sblock.fs_fpg, "%jd");
|
|
if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
CHK(cgp->cg_old_niblk, !=, sblock.fs_ipg, "%jd");
|
|
} else if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
CHK(cgp->cg_niblk, !=, sblock.fs_ipg, "%jd");
|
|
CHK(cgp->cg_initediblk, >, sblock.fs_ipg, "%jd");
|
|
}
|
|
if (cgbase(&sblock, cg) + sblock.fs_fpg < sblock.fs_size) {
|
|
CHK(cgp->cg_ndblk, !=, sblock.fs_fpg, "%jd");
|
|
} else {
|
|
CHK(cgp->cg_ndblk, !=, sblock.fs_size - cgbase(&sblock, cg),
|
|
"%jd");
|
|
}
|
|
start = sizeof(*cgp);
|
|
if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
CHK(cgp->cg_iusedoff, !=, start, "%jd");
|
|
} else if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
CHK(cgp->cg_niblk, !=, 0, "%jd");
|
|
CHK(cgp->cg_initediblk, !=, 0, "%jd");
|
|
CHK(cgp->cg_old_niblk, !=, sblock.fs_ipg, "%jd");
|
|
CHK(cgp->cg_old_btotoff, !=, start, "%jd");
|
|
CHK(cgp->cg_old_boff, !=, cgp->cg_old_btotoff +
|
|
sblock.fs_old_cpg * sizeof(int32_t), "%jd");
|
|
CHK(cgp->cg_iusedoff, !=, cgp->cg_old_boff +
|
|
sblock.fs_old_cpg * sizeof(u_int16_t), "%jd");
|
|
}
|
|
CHK(cgp->cg_freeoff, !=,
|
|
cgp->cg_iusedoff + howmany(sblock.fs_ipg, CHAR_BIT), "%jd");
|
|
if (sblock.fs_contigsumsize == 0) {
|
|
CHK(cgp->cg_nextfreeoff, !=,
|
|
cgp->cg_freeoff + howmany(sblock.fs_fpg, CHAR_BIT), "%jd");
|
|
} else {
|
|
CHK(cgp->cg_nclusterblks, !=, cgp->cg_ndblk / sblock.fs_frag,
|
|
"%jd");
|
|
CHK(cgp->cg_clustersumoff, !=,
|
|
roundup(cgp->cg_freeoff + howmany(sblock.fs_fpg, CHAR_BIT),
|
|
sizeof(u_int32_t)) - sizeof(u_int32_t), "%jd");
|
|
CHK(cgp->cg_clusteroff, !=, cgp->cg_clustersumoff +
|
|
(sblock.fs_contigsumsize + 1) * sizeof(u_int32_t), "%jd");
|
|
CHK(cgp->cg_nextfreeoff, !=, cgp->cg_clusteroff +
|
|
howmany(fragstoblks(&sblock, sblock.fs_fpg), CHAR_BIT),
|
|
"%jd");
|
|
}
|
|
if (error == 0)
|
|
return (1);
|
|
if (prevfailcg == cg)
|
|
return (0);
|
|
prevfailcg = cg;
|
|
pfatal("CYLINDER GROUP %d: INTEGRITY CHECK FAILED", cg);
|
|
printf("\n");
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
rebuild_cg(int cg, struct bufarea *cgbp)
|
|
{
|
|
struct cg *cgp = cgbp->b_un.b_cg;
|
|
long start;
|
|
|
|
/*
|
|
* Zero out the cylinder group and then initialize critical fields.
|
|
* Bit maps and summaries will be recalculated by later passes.
|
|
*/
|
|
memset(cgp, 0, (size_t)sblock.fs_cgsize);
|
|
cgp->cg_magic = CG_MAGIC;
|
|
cgp->cg_cgx = cg;
|
|
cgp->cg_niblk = sblock.fs_ipg;
|
|
cgp->cg_initediblk = MIN(sblock.fs_ipg, 2 * INOPB(&sblock));
|
|
if (cgbase(&sblock, cg) + sblock.fs_fpg < sblock.fs_size)
|
|
cgp->cg_ndblk = sblock.fs_fpg;
|
|
else
|
|
cgp->cg_ndblk = sblock.fs_size - cgbase(&sblock, cg);
|
|
start = sizeof(*cgp);
|
|
if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
cgp->cg_iusedoff = start;
|
|
} else if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
cgp->cg_niblk = 0;
|
|
cgp->cg_initediblk = 0;
|
|
cgp->cg_old_ncyl = sblock.fs_old_cpg;
|
|
cgp->cg_old_niblk = sblock.fs_ipg;
|
|
cgp->cg_old_btotoff = start;
|
|
cgp->cg_old_boff = cgp->cg_old_btotoff +
|
|
sblock.fs_old_cpg * sizeof(int32_t);
|
|
cgp->cg_iusedoff = cgp->cg_old_boff +
|
|
sblock.fs_old_cpg * sizeof(u_int16_t);
|
|
}
|
|
cgp->cg_freeoff = cgp->cg_iusedoff + howmany(sblock.fs_ipg, CHAR_BIT);
|
|
cgp->cg_nextfreeoff = cgp->cg_freeoff + howmany(sblock.fs_fpg,CHAR_BIT);
|
|
if (sblock.fs_contigsumsize > 0) {
|
|
cgp->cg_nclusterblks = cgp->cg_ndblk / sblock.fs_frag;
|
|
cgp->cg_clustersumoff =
|
|
roundup(cgp->cg_nextfreeoff, sizeof(u_int32_t));
|
|
cgp->cg_clustersumoff -= sizeof(u_int32_t);
|
|
cgp->cg_clusteroff = cgp->cg_clustersumoff +
|
|
(sblock.fs_contigsumsize + 1) * sizeof(u_int32_t);
|
|
cgp->cg_nextfreeoff = cgp->cg_clusteroff +
|
|
howmany(fragstoblks(&sblock, sblock.fs_fpg), CHAR_BIT);
|
|
}
|
|
cgp->cg_ckhash = calculate_crc32c(~0L, (void *)cgp, sblock.fs_cgsize);
|
|
cgdirty(cgbp);
|
|
}
|
|
|
|
/*
|
|
* allocate a data block with the specified number of fragments
|
|
*/
|
|
ufs2_daddr_t
|
|
allocblk(long startcg, long frags,
|
|
ufs2_daddr_t (*checkblkavail)(ufs2_daddr_t blkno, long frags))
|
|
{
|
|
ufs2_daddr_t blkno, newblk;
|
|
|
|
if (sujrecovery && checkblkavail == std_checkblkavail) {
|
|
pfatal("allocblk: std_checkblkavail used for SUJ recovery\n");
|
|
return (0);
|
|
}
|
|
if (frags <= 0 || frags > sblock.fs_frag)
|
|
return (0);
|
|
for (blkno = MAX(cgdata(&sblock, startcg), 0);
|
|
blkno < maxfsblock - sblock.fs_frag;
|
|
blkno += sblock.fs_frag) {
|
|
if ((newblk = (*checkblkavail)(blkno, frags)) == 0)
|
|
continue;
|
|
if (newblk > 0)
|
|
return (newblk);
|
|
if (newblk < 0)
|
|
blkno = -newblk;
|
|
}
|
|
for (blkno = MAX(cgdata(&sblock, 0), 0);
|
|
blkno < cgbase(&sblock, startcg) - sblock.fs_frag;
|
|
blkno += sblock.fs_frag) {
|
|
if ((newblk = (*checkblkavail)(blkno, frags)) == 0)
|
|
continue;
|
|
if (newblk > 0)
|
|
return (newblk);
|
|
if (newblk < 0)
|
|
blkno = -newblk;
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
ufs2_daddr_t
|
|
std_checkblkavail(ufs2_daddr_t blkno, long frags)
|
|
{
|
|
struct bufarea *cgbp;
|
|
struct cg *cgp;
|
|
ufs2_daddr_t j, k, baseblk;
|
|
long cg;
|
|
|
|
if ((u_int64_t)blkno > sblock.fs_size)
|
|
return (0);
|
|
for (j = 0; j <= sblock.fs_frag - frags; j++) {
|
|
if (testbmap(blkno + j))
|
|
continue;
|
|
for (k = 1; k < frags; k++)
|
|
if (testbmap(blkno + j + k))
|
|
break;
|
|
if (k < frags) {
|
|
j += k;
|
|
continue;
|
|
}
|
|
cg = dtog(&sblock, blkno + j);
|
|
cgbp = cglookup(cg);
|
|
cgp = cgbp->b_un.b_cg;
|
|
if (!check_cgmagic(cg, cgbp))
|
|
return (-((cg + 1) * sblock.fs_fpg - sblock.fs_frag));
|
|
baseblk = dtogd(&sblock, blkno + j);
|
|
for (k = 0; k < frags; k++) {
|
|
setbmap(blkno + j + k);
|
|
clrbit(cg_blksfree(cgp), baseblk + k);
|
|
}
|
|
n_blks += frags;
|
|
if (frags == sblock.fs_frag)
|
|
cgp->cg_cs.cs_nbfree--;
|
|
else
|
|
cgp->cg_cs.cs_nffree -= frags;
|
|
cgdirty(cgbp);
|
|
return (blkno + j);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Check whether a file size is within the limits for the filesystem.
|
|
* Return 1 when valid and 0 when too big.
|
|
*
|
|
* This should match the file size limit in ffs_mountfs().
|
|
*/
|
|
int
|
|
chkfilesize(mode_t mode, u_int64_t filesize)
|
|
{
|
|
u_int64_t kernmaxfilesize;
|
|
|
|
if (sblock.fs_magic == FS_UFS1_MAGIC)
|
|
kernmaxfilesize = (off_t)0x40000000 * sblock.fs_bsize - 1;
|
|
else
|
|
kernmaxfilesize = sblock.fs_maxfilesize;
|
|
if (filesize > kernmaxfilesize ||
|
|
filesize > sblock.fs_maxfilesize ||
|
|
(mode == IFDIR && filesize > MAXDIRSIZE)) {
|
|
if (debug)
|
|
printf("bad file size %ju:", (uintmax_t)filesize);
|
|
return (0);
|
|
}
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* Slow down IO so as to leave some disk bandwidth for other processes
|
|
*/
|
|
void
|
|
slowio_start()
|
|
{
|
|
|
|
/* Delay one in every 8 operations */
|
|
slowio_pollcnt = (slowio_pollcnt + 1) & 7;
|
|
if (slowio_pollcnt == 0) {
|
|
gettimeofday(&slowio_starttime, NULL);
|
|
}
|
|
}
|
|
|
|
void
|
|
slowio_end()
|
|
{
|
|
struct timeval tv;
|
|
int delay_usec;
|
|
|
|
if (slowio_pollcnt != 0)
|
|
return;
|
|
|
|
/* Update the slowdown interval. */
|
|
gettimeofday(&tv, NULL);
|
|
delay_usec = (tv.tv_sec - slowio_starttime.tv_sec) * 1000000 +
|
|
(tv.tv_usec - slowio_starttime.tv_usec);
|
|
if (delay_usec < 64)
|
|
delay_usec = 64;
|
|
if (delay_usec > 2500000)
|
|
delay_usec = 2500000;
|
|
slowio_delay_usec = (slowio_delay_usec * 63 + delay_usec) >> 6;
|
|
/* delay by 8 times the average IO delay */
|
|
if (slowio_delay_usec > 64)
|
|
usleep(slowio_delay_usec * 8);
|
|
}
|
|
|
|
/*
|
|
* Find a pathname
|
|
*/
|
|
void
|
|
getpathname(char *namebuf, ino_t curdir, ino_t ino)
|
|
{
|
|
int len;
|
|
char *cp;
|
|
struct inode ip;
|
|
struct inodesc idesc;
|
|
static int busy = 0;
|
|
|
|
if (curdir == ino && ino == UFS_ROOTINO) {
|
|
(void)strcpy(namebuf, "/");
|
|
return;
|
|
}
|
|
if (busy || !INO_IS_DVALID(curdir)) {
|
|
(void)strcpy(namebuf, "?");
|
|
return;
|
|
}
|
|
busy = 1;
|
|
memset(&idesc, 0, sizeof(struct inodesc));
|
|
idesc.id_type = DATA;
|
|
idesc.id_fix = IGNORE;
|
|
cp = &namebuf[MAXPATHLEN - 1];
|
|
*cp = '\0';
|
|
if (curdir != ino) {
|
|
idesc.id_parent = curdir;
|
|
goto namelookup;
|
|
}
|
|
while (ino != UFS_ROOTINO) {
|
|
idesc.id_number = ino;
|
|
idesc.id_func = findino;
|
|
idesc.id_name = strdup("..");
|
|
ginode(ino, &ip);
|
|
if ((ckinode(ip.i_dp, &idesc) & FOUND) == 0) {
|
|
irelse(&ip);
|
|
free(idesc.id_name);
|
|
break;
|
|
}
|
|
irelse(&ip);
|
|
free(idesc.id_name);
|
|
namelookup:
|
|
idesc.id_number = idesc.id_parent;
|
|
idesc.id_parent = ino;
|
|
idesc.id_func = findname;
|
|
idesc.id_name = namebuf;
|
|
ginode(idesc.id_number, &ip);
|
|
if ((ckinode(ip.i_dp, &idesc) & FOUND) == 0) {
|
|
irelse(&ip);
|
|
break;
|
|
}
|
|
irelse(&ip);
|
|
len = strlen(namebuf);
|
|
cp -= len;
|
|
memmove(cp, namebuf, (size_t)len);
|
|
*--cp = '/';
|
|
if (cp < &namebuf[UFS_MAXNAMLEN])
|
|
break;
|
|
ino = idesc.id_number;
|
|
}
|
|
busy = 0;
|
|
if (ino != UFS_ROOTINO)
|
|
*--cp = '?';
|
|
memmove(namebuf, cp, (size_t)(&namebuf[MAXPATHLEN] - cp));
|
|
}
|
|
|
|
void
|
|
catch(int sig __unused)
|
|
{
|
|
|
|
ckfini(0);
|
|
exit(12);
|
|
}
|
|
|
|
/*
|
|
* When preening, allow a single quit to signal
|
|
* a special exit after file system checks complete
|
|
* so that reboot sequence may be interrupted.
|
|
*/
|
|
void
|
|
catchquit(int sig __unused)
|
|
{
|
|
printf("returning to single-user after file system check\n");
|
|
returntosingle = 1;
|
|
(void)signal(SIGQUIT, SIG_DFL);
|
|
}
|
|
|
|
/*
|
|
* determine whether an inode should be fixed.
|
|
*/
|
|
int
|
|
dofix(struct inodesc *idesc, const char *msg)
|
|
{
|
|
|
|
switch (idesc->id_fix) {
|
|
|
|
case DONTKNOW:
|
|
if (idesc->id_type == DATA)
|
|
direrror(idesc->id_number, msg);
|
|
else
|
|
pwarn("%s", msg);
|
|
if (preen) {
|
|
printf(" (SALVAGED)\n");
|
|
idesc->id_fix = FIX;
|
|
return (ALTERED);
|
|
}
|
|
if (reply("SALVAGE") == 0) {
|
|
idesc->id_fix = NOFIX;
|
|
return (0);
|
|
}
|
|
idesc->id_fix = FIX;
|
|
return (ALTERED);
|
|
|
|
case FIX:
|
|
return (ALTERED);
|
|
|
|
case NOFIX:
|
|
case IGNORE:
|
|
return (0);
|
|
|
|
default:
|
|
errx(EEXIT, "UNKNOWN INODESC FIX MODE %d", idesc->id_fix);
|
|
}
|
|
/* NOTREACHED */
|
|
return (0);
|
|
}
|
|
|
|
#include <stdarg.h>
|
|
|
|
/*
|
|
* Print details about a buffer.
|
|
*/
|
|
void
|
|
prtbuf(struct bufarea *bp, const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (preen)
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
printf(": bp %p, type %s, bno %jd, size %d, refcnt %d, flags %s, "
|
|
"index %jd\n", bp, BT_BUFTYPE(bp->b_type), (intmax_t) bp->b_bno,
|
|
bp->b_size, bp->b_refcnt, bp->b_flags & B_DIRTY ? "dirty" : "clean",
|
|
(intmax_t) bp->b_index);
|
|
}
|
|
|
|
/*
|
|
* An unexpected inconsistency occurred.
|
|
* Die if preening or file system is running with soft dependency protocol,
|
|
* otherwise just print message and continue.
|
|
*/
|
|
void
|
|
pfatal(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (!preen) {
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
if (usedsoftdep)
|
|
(void)fprintf(stdout,
|
|
"\nUNEXPECTED SOFT UPDATE INCONSISTENCY\n");
|
|
/*
|
|
* Force foreground fsck to clean up inconsistency.
|
|
*/
|
|
if (bkgrdflag) {
|
|
cmd.value = FS_NEEDSFSCK;
|
|
cmd.size = 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FS_NEEDSFSCK FLAG\n");
|
|
fprintf(stdout, "CANNOT RUN IN BACKGROUND\n");
|
|
ckfini(0);
|
|
exit(EEXIT);
|
|
}
|
|
return;
|
|
}
|
|
if (cdevname == NULL)
|
|
cdevname = strdup("fsck");
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
(void)fprintf(stdout,
|
|
"\n%s: UNEXPECTED%sINCONSISTENCY; RUN fsck MANUALLY.\n",
|
|
cdevname, usedsoftdep ? " SOFT UPDATE " : " ");
|
|
/*
|
|
* Force foreground fsck to clean up inconsistency.
|
|
*/
|
|
if (bkgrdflag) {
|
|
cmd.value = FS_NEEDSFSCK;
|
|
cmd.size = 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FS_NEEDSFSCK FLAG\n");
|
|
}
|
|
ckfini(0);
|
|
exit(EEXIT);
|
|
}
|
|
|
|
/*
|
|
* Pwarn just prints a message when not preening or running soft dependency
|
|
* protocol, or a warning (preceded by filename) when preening.
|
|
*/
|
|
void
|
|
pwarn(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (preen)
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
}
|
|
|
|
/*
|
|
* Stub for routines from kernel.
|
|
*/
|
|
void
|
|
panic(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
pfatal("INTERNAL INCONSISTENCY:");
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
exit(EEXIT);
|
|
}
|