mirror of
https://git.hardenedbsd.org/hardenedbsd/HardenedBSD.git
synced 2024-11-23 04:11:07 +01:00
df57947f08
The Software Package Data Exchange (SPDX) group provides a specification to make it easier for automated tools to detect and summarize well known opensource licenses. We are gradually adopting the specification, noting that the tags are considered only advisory and do not, in any way, superceed or replace the license texts. Special thanks to Wind River for providing access to "The Duke of Highlander" tool: an older (2014) run over FreeBSD tree was useful as a starting point. Initially, only tag files that use BSD 4-Clause "Original" license. RelNotes: yes Differential Revision: https://reviews.freebsd.org/D13133
369 lines
7.7 KiB
Bash
369 lines
7.7 KiB
Bash
#!/bin/sh
|
|
#
|
|
# SPDX-License-Identifier: BSD-4-Clause
|
|
#
|
|
# Copyright (c) 1994 Geoffrey M. Rehmet, Rhodes University
|
|
# All rights reserved.
|
|
#
|
|
# Redistribution and use in source and binary forms, with or without
|
|
# modification, are permitted provided that the following conditions
|
|
# are met:
|
|
# 1. Redistributions of source code must retain the above copyright
|
|
# notice, this list of conditions and the following disclaimer.
|
|
# 2. Redistributions in binary form must reproduce the above copyright
|
|
# notice, this list of conditions and the following disclaimer in the
|
|
# documentation and/or other materials provided with the distribution.
|
|
# 3. All advertising materials mentioning features or use of this software
|
|
# must display the following acknowledgement:
|
|
# This product includes software developed by Geoffrey M. Rehmet
|
|
# 4. Neither the name of Geoffrey M. Rehmet nor that of Rhodes University
|
|
# may be used to endorse or promote products derived from this software
|
|
# without specific prior written permission.
|
|
#
|
|
# THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED
|
|
# WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
# MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
# IN NO EVENT SHALL GEOFFREY M. REHMET OR RHODES UNIVERSITY BE LIABLE
|
|
# FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
# DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
# OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
# HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
# OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
# SUCH DAMAGE.
|
|
#
|
|
# $FreeBSD$
|
|
#
|
|
# manctl:
|
|
# a utility for manipulating manual pages
|
|
# functions:
|
|
# compress uncompressed man pages (elliminating .so's)
|
|
# this is now two-pass. If possible, .so's
|
|
# are replaced with hard links
|
|
# uncompress compressed man pages
|
|
# Things to watch out for:
|
|
# Hard links - careful with g(un)zipping!
|
|
# .so's - throw everything through soelim before gzip!
|
|
# symlinks - ignore these - eg: expn is its own man page:
|
|
# don't want to compress this!
|
|
#
|
|
PATH=/bin:/sbin:/usr/bin:/usr/sbin; export PATH
|
|
|
|
#
|
|
# Uncompress one page
|
|
#
|
|
uncompress_page()
|
|
{
|
|
local pname
|
|
local fname
|
|
local sect
|
|
local ext
|
|
|
|
# break up file name
|
|
pname=$1
|
|
IFS='.' ; set $pname
|
|
# less than 3 fields - don't know what to do with this
|
|
if [ $# -lt 3 ] ; then
|
|
IFS=" " ; echo ignoring $pname 1>&2 ; return 0 ;
|
|
fi
|
|
# construct name and section
|
|
fname=$1 ; shift
|
|
while [ $# -gt 2 ] ; do
|
|
fname=$fname.$1
|
|
shift
|
|
done
|
|
sect=$1
|
|
ext=$2
|
|
|
|
IFS=" "
|
|
case "$ext" in
|
|
gz|Z) {
|
|
IFS=" " ; set `file $pname`
|
|
if [ $2 != "gzip" ] ; then
|
|
echo moving hard link $pname 1>&2
|
|
mv $pname $fname.$ext # link
|
|
else
|
|
if [ $2 != "symbolic" ] ; then
|
|
echo gunzipping page $pname 1>&2
|
|
temp=`mktemp -t manager` || exit 1
|
|
gunzip -c $pname > $temp
|
|
chmod u+w $pname
|
|
cp $temp $pname
|
|
chmod 444 $pname
|
|
mv $pname $fname.$sect
|
|
rm -f $temp
|
|
else
|
|
# skip symlinks - this can be
|
|
# a program like expn, which is
|
|
# its own man page !
|
|
echo skipping symlink $pname 1>&2
|
|
fi
|
|
fi };;
|
|
*) {
|
|
IFS=" "
|
|
echo skipping file $pname 1>&2
|
|
} ;;
|
|
esac
|
|
# reset IFS - this is important!
|
|
IFS=" "
|
|
}
|
|
|
|
|
|
#
|
|
# Uncompress manpages in paths
|
|
#
|
|
do_uncompress()
|
|
{
|
|
local i
|
|
local dir
|
|
local workdir
|
|
|
|
workdir=`pwd`
|
|
while [ $# != 0 ] ; do
|
|
if [ -d $1 ] ; then
|
|
dir=$1
|
|
cd $dir
|
|
for i in * ; do
|
|
case $i in
|
|
*cat?) ;; # ignore cat directories
|
|
*) {
|
|
if [ -d $i ] ; then
|
|
do_uncompress $i
|
|
else
|
|
if [ -e $i ] ; then
|
|
uncompress_page $i
|
|
fi
|
|
fi } ;;
|
|
esac
|
|
done
|
|
cd $workdir
|
|
else
|
|
echo "directory $1 not found" 1>&2
|
|
fi
|
|
shift
|
|
done
|
|
}
|
|
|
|
#
|
|
# Remove .so's from one file
|
|
#
|
|
so_purge_page()
|
|
{
|
|
local so_entries
|
|
local lines
|
|
local fname
|
|
|
|
so_entries=`grep "^\.so" $1 | wc -l`
|
|
if [ $so_entries -eq 0 ] ; then return 0 ; fi
|
|
|
|
# we have a page with a .so in it
|
|
echo $1 contains a .so entry 2>&1
|
|
|
|
# now check how many lines in the file
|
|
lines=`wc -l < $1`
|
|
|
|
# if the file is only one line long, we can replace it
|
|
# with a hard link!
|
|
if [ $lines -eq 1 ] ; then
|
|
fname=$1;
|
|
echo replacing $fname with a hard link
|
|
set `cat $fname`;
|
|
rm -f $fname
|
|
ln ../$2 $fname
|
|
else
|
|
echo inlining page $fname 1>&2
|
|
temp=`mktemp -t manager` || exit 1
|
|
cat $fname | \
|
|
(cd .. ; soelim ) > $temp
|
|
chmod u+w $fname
|
|
cp $temp $fname
|
|
chmod 444 $fname
|
|
rm -f $temp
|
|
fi
|
|
}
|
|
|
|
#
|
|
# Remove .so entries from man pages
|
|
# If a page consists of just one line with a .so,
|
|
# replace it with a hard link
|
|
#
|
|
remove_so()
|
|
{
|
|
local pname
|
|
local fname
|
|
local sect
|
|
|
|
# break up file name
|
|
pname=$1
|
|
IFS='.' ; set $pname
|
|
if [ $# -lt 2 ] ; then
|
|
IFS=" " ; echo ignoring $pname 1>&2 ; return 0 ;
|
|
fi
|
|
# construct name and section
|
|
fname=$1 ; shift
|
|
while [ $# -gt 1 ] ; do
|
|
fname=$fname.$1
|
|
shift
|
|
done
|
|
sect=$1
|
|
|
|
IFS=" "
|
|
case "$sect" in
|
|
gz) { echo file $pname already gzipped 1>&2 ; } ;;
|
|
Z) { echo file $pname already compressed 1>&2 ; } ;;
|
|
[12345678ln]*){
|
|
IFS=" " ; set `file $pname`
|
|
if [ $2 = "gzip" ] ; then
|
|
echo moving hard link $pname 1>&2
|
|
mv $pname $pname.gz # link
|
|
else
|
|
if [ $2 != "symbolic" ] ; then
|
|
echo "removing .so's in page $pname" 1>&2
|
|
so_purge_page $pname
|
|
else
|
|
# skip symlink - this can be
|
|
# a program like expn, which is
|
|
# its own man page !
|
|
echo skipping symlink $pname 1>&2
|
|
fi
|
|
fi };;
|
|
*) {
|
|
IFS=" "
|
|
echo skipping file $pname 1>&2
|
|
} ;;
|
|
esac
|
|
# reset IFS - this is important!
|
|
IFS=" "
|
|
}
|
|
|
|
|
|
#
|
|
# compress one page
|
|
# We need to watch out for hard links here.
|
|
#
|
|
compress_page()
|
|
{
|
|
local pname
|
|
local fname
|
|
local sect
|
|
|
|
# break up file name
|
|
pname=$1
|
|
IFS='.' ; set $pname
|
|
if [ $# -lt 2 ] ; then
|
|
IFS=" " ; echo ignoring $pname 1>&2 ; return 0 ;
|
|
fi
|
|
# construct name and section
|
|
fname=$1 ; shift
|
|
while [ $# -gt 1 ] ; do
|
|
fname=$fname.$1
|
|
shift
|
|
done
|
|
sect=$1
|
|
|
|
IFS=" "
|
|
case "$sect" in
|
|
gz) { echo file $pname already gzipped 1>&2 ; } ;;
|
|
Z) { echo file $pname already compressed 1>&2 ; } ;;
|
|
[12345678ln]*){
|
|
IFS=" " ; set `file $pname`
|
|
if [ $2 = "gzip" ] ; then
|
|
echo moving hard link $pname 1>&2
|
|
mv $pname $pname.gz # link
|
|
else
|
|
if [ $2 != "symbolic" ] ; then
|
|
echo gzipping page $pname 1>&2
|
|
temp=`mktemp -t manager` || exit 1
|
|
cat $pname | \
|
|
(cd .. ; soelim )| gzip -c -- > $temp
|
|
chmod u+w $pname
|
|
cp $temp $pname
|
|
chmod 444 $pname
|
|
mv $pname $pname.gz
|
|
rm -f $temp
|
|
else
|
|
# skip symlink - this can be
|
|
# a program like expn, which is
|
|
# its own man page !
|
|
echo skipping symlink $pname 1>&2
|
|
fi
|
|
fi };;
|
|
*) {
|
|
IFS=" "
|
|
echo skipping file $pname 1>&2
|
|
} ;;
|
|
esac
|
|
# reset IFS - this is important!
|
|
IFS=" "
|
|
}
|
|
|
|
#
|
|
# Compress man pages in paths
|
|
#
|
|
do_compress_so()
|
|
{
|
|
local i
|
|
local dir
|
|
local workdir
|
|
local what
|
|
|
|
what=$1
|
|
shift
|
|
workdir=`pwd`
|
|
while [ $# != 0 ] ; do
|
|
if [ -d $1 ] ; then
|
|
dir=$1
|
|
cd $dir
|
|
for i in * ; do
|
|
case $i in
|
|
*cat?) ;; # ignore cat directories
|
|
*) {
|
|
if [ -d $i ] ; then
|
|
do_compress_so $what $i
|
|
else
|
|
if [ -e $i ] ; then
|
|
$what $i
|
|
fi
|
|
fi } ;;
|
|
esac
|
|
done
|
|
cd $workdir
|
|
else
|
|
echo "directory $1 not found" 1>&2
|
|
fi
|
|
shift
|
|
done
|
|
}
|
|
|
|
#
|
|
# Display a usage message
|
|
#
|
|
ctl_usage()
|
|
{
|
|
echo "usage: $1 -compress <path> ... " 1>&2
|
|
echo " $1 -uncompress <path> ... " 1>&2
|
|
exit 1
|
|
}
|
|
|
|
#
|
|
# remove .so's and do compress
|
|
#
|
|
do_compress()
|
|
{
|
|
# First remove all so's from the pages to be compressed
|
|
do_compress_so remove_so "$@"
|
|
# now do ahead and compress the pages
|
|
do_compress_so compress_page "$@"
|
|
}
|
|
|
|
#
|
|
# dispatch options
|
|
#
|
|
if [ $# -lt 2 ] ; then ctl_usage $0 ; fi ;
|
|
|
|
case "$1" in
|
|
-compress) shift ; do_compress "$@" ;;
|
|
-uncompress) shift ; do_uncompress "$@" ;;
|
|
*) ctl_usage $0 ;;
|
|
esac
|