diff --git a/ancient/README.md b/ancient/README.md new file mode 100644 index 0000000..6b17961 --- /dev/null +++ b/ancient/README.md @@ -0,0 +1,5 @@ +Only kept for historic reasons: + +- use at your own risk +- use for inspiration +- but don't use for production diff --git a/qos-neu b/ancient/qos-neu similarity index 100% rename from qos-neu rename to ancient/qos-neu diff --git a/statistics b/statistics deleted file mode 100755 index b5751a1..0000000 --- a/statistics +++ /dev/null @@ -1,571 +0,0 @@ -#!/usr/bin/perl -wT -# -# TWiki Collaboration Platform, http://TWiki.org/ -# -# Copyright (C) 2000-2003 Peter Thoeny, peter@thoeny.com -# Copyright (C) 2002 Richard Donkin, rdonkin@bigfoot.com -# -# For licensing info read license.txt file in the TWiki root. -# This program is free software; you can redistribute it and/or -# modify it under the terms of the GNU General Public License -# as published by the Free Software Foundation; either version 2 -# of the License, or (at your option) any later version. -# -# This program is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details, published at -# http://www.gnu.org/copyleft/gpl.html -# -# DESCRIPTION -# This script does update the usage statistics of each TWiki web. -# It reads the current month's log file and updates the table -# in the WebStatistics topic of each web. -# The script should be called by a cron job, it is recommended to -# call it once a day. - -# Set library paths in @INC, at compile time -BEGIN { unshift @INC, '.'; require 'setlib.cfg'; } - -# 'Use locale' for internationalisation of Perl regexes - -# main locale settings are done in TWiki::setupLocale -BEGIN { - # Do a dynamic 'use locale' for this module - if( $TWiki::useLocale ) { - require locale; - import locale (); - } -} - -use CGI::Carp qw(fatalsToBrowser); -use CGI; -use File::Copy qw(copy); -use IO::File; - -use TWiki; - -use strict; -# use diagnostics; # Debug only - -#open(STDERR,'>&STDOUT'); # redirect error to browser -# FIXME: Beware global $| in mod_perl! Should use 'local' -$| = 1; # no buffering - -&main(); - -# ========================= -sub main -{ - # initialize variables - my $cgiQuery = undef; - my $tmp = ""; - my $theTopic = ""; - my $thePathInfo = ""; - my $theRemoteUser = ""; - my $logDate = ""; - - # determine at runtime if script is called by browser or cron job - if( $ENV{'DOCUMENT_ROOT'} ) { - # script is called by browser - $cgiQuery = new CGI; - $thePathInfo = $cgiQuery->path_info() || ""; - $theRemoteUser = $cgiQuery->remote_user() || ""; - $theTopic = $cgiQuery->param( 'topic' ) || ""; - $tmp = $cgiQuery->param( 'logdate' ) || ""; - $tmp =~ s/[^0-9]//g; # remove all non numerals - if( $tmp ne "" ) { - $logDate = "$tmp"; - } - TWiki::writeHeader( $cgiQuery ); - print "\n\nTWiki: Create Usage Statistics\n"; - print "\n\n"; - } else { - # script is called by cron job - } - - # Set up locale and regexes - &TWiki::basicInitialize(); - - # Initial messages - printMsg( "TWiki: Create Usage Statistics", $cgiQuery ); - if( $cgiQuery ) { - print "

Do not interrupt this script! ( Please wait until page download has finished )

\n"; - } - if ( $theRemoteUser ) { - $tmp = &TWiki::userToWikiName( $theRemoteUser ); - $tmp =~ s/^Main\.//; - printMsg( "* Executed by $tmp", $cgiQuery ); - } else { - printMsg( "* Executed by a guest or a cron job scheduler", $cgiQuery ); - } - - if( ! $logDate ) { - # get current local time and format to "yyyymm" format: - my ( $sec, $min, $hour, $mday, $mon, $year) = localtime( time() ); - $year = sprintf("%.4u", $year + 1900); # Y2K fix - $mon = $mon+1; - $logDate = sprintf("%.4u%.2u", $year, $mon); - } - - my $logMonth; - my $logYear; - $tmp = $logDate; - $tmp =~ s/([0-9]{4})(.*)/$2/g; - if( $tmp && $tmp < 13 ) { - $logMonth = $TWiki::isoMonth[$tmp-1]; - } else { - $logMonth = "Date error"; - } - $logYear = $logDate; - $logYear =~ s/([0-9]{4})(.*)/$1/g; - my $logMonthYear = "$logMonth $logYear"; - printMsg( "* Statistics for $logMonthYear", $cgiQuery ); - - my $logFile = $TWiki::logFilename; - $logFile =~ s/%DATE%/$logDate/g; - - if( -e $logFile ) { - # Copy the log file to temp file, since analysis could take some time - - # FIXME move the temp dir stuff to TWiki.cfg - my $tmpDir; - if ( $TWiki::OS eq "UNIX" ) { - $tmpDir = $ENV{'TEMP'} || "/tmp"; - } elsif ( $TWiki::OS eq "WINDOWS" ) { - $tmpDir = $ENV{'TEMP'} || "c:/"; - } else { - # FIXME handle other OSs properly - assume Unix for now. - $tmpDir = "/tmp"; - } - my $randNo = int ( rand 1000); # For mod_perl with threading... - my $tmpFilename = "$tmpDir/twiki-stats.$$.$randNo"; - $tmpFilename =~ /(.*)/; $tmpFilename = $1; # Untaint - - File::Copy::copy ($logFile, $tmpFilename) - or die "Can't copy $logFile to $tmpFilename - $!"; - - # Open the temp file - my $TMPFILE = new IO::File; - open $TMPFILE, $tmpFilename - or die "Can't open $tmpFilename - $!"; - - # Do a single data collection pass on the temporary copy of logfile, - # then call processWeb once for each web of interest. - my ($viewRef, $contribRef, $statViewsRef, $statSavesRef, - $statUploadsRef) = collectLogData( $TMPFILE, $logMonthYear ); - -=pod - # DEBUG ONLY - debugPrintHash($viewRef); - debugPrintHash($contribRef); - print "statViews tests===========\n"; - print "Views in Main = " . ${$statViewsRef}{'Main'} . "\n"; - print "hash stats (used/avail) = " . %{$statViewsRef}."\n"; - - foreach my $web (keys %{$statViewsRef}) { - print "Web summary for $web\n"; - print $statViewsRef->{$web}."\n"; - print $statSavesRef->{$web}."\n"; - print $statUploadsRef->{$web}."\n"; - } -=cut - - # Generate WebStatistics topic update for one or more webs - if ( $thePathInfo ) { - # do a particular web: - processWeb( $thePathInfo, $theRemoteUser, $theTopic, $logMonthYear, $viewRef, $contribRef, - $statViewsRef, $statSavesRef, $statUploadsRef, $cgiQuery ); - - } else { - # do all webs: - my $dataDir = &TWiki::getDataDir(); - my @weblist = (); - if( opendir( DIR, "$dataDir" ) ) { - @weblist = grep /^[^\.\_]/, readdir DIR; # exclude webs starting with . or _ - closedir DIR; - } else { - printMsg( " *** Error: opendir $dataDir, $!", $cgiQuery ); - } - foreach my $web ( @weblist ) { - if( -d "$dataDir/$web" ) { - processWeb( "/$web", $theRemoteUser, $theTopic, $logMonthYear, $viewRef, $contribRef, - $statViewsRef, $statSavesRef, $statUploadsRef, $cgiQuery ); - } - } - } - close $TMPFILE; # Shouldn't be necessary with 'my' - unlink $tmpFilename; # FIXME: works on Windows??? Unlink before - # usage to ensure deleted on crash? - } else { - printMsg( " - Note: Log file $logFile does not exist", $cgiQuery ); - } - - if( $cgiQuery ) { - $tmp = $TWiki::statisticsTopicname; - my $url = &TWiki::getViewUrl( "", $tmp ); - printMsg( "* Go back to $tmp topic", $cgiQuery ); - printMsg( "End creating usage statistics", $cgiQuery ); - print "\n"; - } else { - printMsg( "End creating usage statistics", $cgiQuery ); - } -} - -# Debug only -# Print all entries in a view or contrib hash, sorted by web and item name -sub debugPrintHash { - my ($statsRef) = @_; - # print "Main.WebHome views = " . ${$statsRef}{'Main'}{'WebHome'}."\n"; - # print "Main web, TWikiGuest contribs = " . ${$statsRef}{'Main'}{'Main.TWikiGuest'}."\n"; - foreach my $web ( sort keys %$statsRef) { - my $count = 0; - print "$web web:\n"; - # Get reference to the sub-hash for this web - my $webhashref = ${$statsRef}{$web}; - # print "webhashref is " . ref ($webhashref) ."\n"; - # Items can be topics (for view hash) or users (for contrib hash) - foreach my $item ( sort keys %$webhashref ) { - print " $item = "; - print "" . ( ${$webhashref}{$item} || 0 ) ."\n"; - $count += ${$webhashref}{$item}; - } - print " WEB TOTAL = $count\n"; - } -} - - -# ========================= -# Process the whole log file and collect information in hash tables. -# Must build stats for all webs, to handle case of renames into web -# requested for a single-web statistics run. -# -# Main hash tables are divided by web: -# -# $view{$web}{$TopicName} == number of views, by topic -# $contrib{$web}{"Main.".$WikiName} == number of saves/uploads, by user - -sub collectLogData -{ - my( $TMPFILE, $theLogMonthYear ) = @_; - - # Examples of log file format: - # | 03 Feb 2000 - 02:43 | Main.PeterThoeny | view | Know.WebHome | | - # | 03 Feb 2000 - 02:43 | Main.PeterThoeny | save | Know.WebHome | | - # | 03 Feb 2000 - 02:53 | Main.PeterThoeny | save | Know.WebHome | repRev 1.7 Main.PeterThoeny 2000/02/03 02:43:22 | - # | 23 Feb 2002 - 11:07 | Main.TWikiGuest | search | Main | Office *Locations[^A-Za-z] | 127.0.0.1 | - # Note: there's no topic name on search log entry - # | 23 Feb 2002 - 11:07 | Main.guest | search | Main | Office *Locations[^A-Za-z] | 127.0.0.1 | - # | 28 Mar 2002 - 07:11 | Main.FredBloggs | rename | Test.TestTopic7 | moved to Test.TestTopic7New | 127.0.0.1 | - - - my %view; # Hash of hashes, counts topic views by - my %contrib; # Hash of hashes, counts uploads/saves by - - # Hashes for each type of statistic, one hash entry per web - my %statViews; - my %statSaves; - my %statUploads; - - # Imported regex objects, supporting I18N - my $webNameRegex = $TWiki::webNameRegex; - my $wikiWordRegex = $TWiki::wikiWordRegex; - my $abbrevRegex = $TWiki::abbrevRegex; - - # Script regexes - my $intranetUserRegex = qr/[a-z0-9]+/; # FIXME: should centralise this - my $userRegex = qr/(?:$intranetUserRegex|$wikiWordRegex)/; - my $opRegex = qr/[a-z0-9]+/; # Operation, no i18n needed - # my $topicRegex = qr/(?:$wikiWordRegex|$abbrevRegex)/; # Strict topic names only - my $topicRegex = qr/[^ ]+/; # Relaxed topic names - any non-space OK - # but won't be auto-linked in WebStatistics - my $errorRegex = qr/\(not exist\)/; # Match '(not exist)' flag - - my ($webName, $opName, $topicName, $userName, $newTopicName, $newTopicWeb); - binmode $TMPFILE; - while ( <$TMPFILE> ) { - my $line = $_; - $line =~ s/\r*\n$//; # Clean out line endings - - $line =~ /^\|[^\|]*\| ($webNameRegex\.$userRegex) \| ($opRegex) \| ($webNameRegex)[. ]/o; - $userName = $1 || ""; # Main.FredBloggs - $opName = $2 || ""; - $webName = $3 || ""; - - # Skip bad logfile lines and warn if necessary - unless ($userName && $opName && $webName) { - if( $TWiki::doDebugStatistics ) { - TWiki::writeDebug("Invalid log file line = '$line'"); - TWiki::writeDebug("userName = '$userName'"); - TWiki::writeDebug("opName = '$opName'"); - TWiki::writeDebug("webName = '$webName'"); - } - next; - } - - my $logContrib = 0; - - if ($opName eq 'view' ) { - $statViews{$webName}++; - # Pick up the topic name and any error string - $line =~ /^\|[^\|]*\| ($webNameRegex\.$userRegex) \| ($opRegex) \| ($webNameRegex)\.($topicRegex) \| +(${errorRegex}?) */o; - $topicName = $4 || ""; - my $noSuchTopic = $5 || ""; # Set if '(not exist)' matched - - unless( $topicName ) { - if( $TWiki::doDebugStatistics ) { - TWiki::writeDebug("Invalid log file line = '$line'"); - TWiki::writeDebug("userName = '$userName'"); - TWiki::writeDebug("opName = '$opName'"); - TWiki::writeDebug("webName = '$webName'"); - TWiki::writeDebug("topicName = '$topicName'"); - } - next; - } - - # Skip accesses to non-existent topics - if ($noSuchTopic) { - next; - } else { - # Count this topic access - $view{$webName}{$topicName}++; - } - - } elsif ($opName eq 'save' ) { - $statSaves{$webName}++; - $logContrib = 1; - - } elsif ($opName eq 'upload' ) { - $statUploads{$webName}++; - $logContrib = 1; - - } elsif ($opName eq 'rename' ) { - # Pick up the old and new topic names - $line =~ /^\|[^\|]*\| ($webNameRegex\.$userRegex) \| ($opRegex) \| ($webNameRegex)\.($topicRegex) \| moved to ($webNameRegex)\.($topicRegex) /o; - $topicName = $4 || ""; - $newTopicWeb = $5 || ""; - $newTopicName = $6 || ""; - ## TWiki::writeDebug("$topicName renamed to $newTopicWeb.$newTopicName"); - - unless ($topicName && $newTopicWeb && $newTopicName) { - if( $TWiki::doDebugStatistics ) { - TWiki::writeDebug("Invalid log file line (rename) = '$line'"); - TWiki::writeDebug("userName = '$userName'"); - TWiki::writeDebug("opName = '$opName'"); - TWiki::writeDebug("webName = '$webName'"); - TWiki::writeDebug("topicName= '$topicName'"); - TWiki::writeDebug("newTopicWeb= '$newTopicWeb'"); - TWiki::writeDebug("newTopicName = '$newTopicName'"); - } - next; - } - # Get number of views for old topic this month (may be zero) - my $oldViews = $view{$webName}{$topicName} || 0; - - # Transfer views from old to new topic - $view{$newTopicWeb}{$newTopicName} = $oldViews; - delete $view{$webName}{$topicName}; - - # Transfer views from old to new web - if ( $newTopicWeb ne $webName ) { - $statViews{$webName} -= $oldViews; - $statViews{$newTopicWeb} += $oldViews; - } - } - # Record saves and uploads - if ($logContrib) { - # Record the contribution by user name - $contrib{$webName}{$userName}++; - } -=pod - # DEBUG - $. <= 5 && print "$line\n"; - print "$line\n"; - print "$.: $userName did $opName on $webName"; - print ".$topicName" if (defined $topicName); - print "\n"; -=cut - - } - -=pod - - print "Main.WebHome views = " . $view{'Main'}{'WebHome'}."\n"; - print "Main web's contribs = " . $contrib{'Main'}{'Main.RichardDonkin'}."\n"; - debugPrintHash(\%view); - debugPrintHash(\%contrib); -=cut - return \%view, \%contrib, \%statViews, \%statSaves, \%statUploads; - - - - -} - -# ========================= -sub processWeb -{ - my( $thePathInfo, $theRemoteUser, $theTopic, $theLogMonthYear, $viewRef, $contribRef, - $statViewsRef, $statSavesRef, $statUploadsRef, $cgiQuery ) = @_; - - my ( $topic, $webName, $dummy, $userName, $dataDir ) = - &TWiki::initialize( $thePathInfo, $theRemoteUser, $theTopic, "", $cgiQuery ); - $dummy = ""; # to suppress warning - - printMsg( "* Reporting on TWiki.$webName web", $cgiQuery ); - - if( ! &TWiki::Store::webExists( $webName ) ) { - printMsg( " *** Error: Web $webName does not exist", $cgiQuery ); - return; - } - - # Handle null values, print summary message to browser/stdout - my $statViews = $statViewsRef->{$webName}; - my $statSaves = $statSavesRef->{$webName}; - my $statUploads = $statUploadsRef->{$webName}; - $statViews ||= 0; - $statSaves ||= 0; - $statUploads ||= 0; - printMsg( " - view: $statViews, save: $statSaves, upload: $statUploads", $cgiQuery ); - - - # Get the top N views and contribs in this web - my (@topViews) = getTopList( $TWiki::statsTopViews, $webName, $viewRef ); - my (@topContribs) = getTopList( $TWiki::statsTopContrib, $webName, $contribRef ); - - # Print information to stdout - my $statTopViews = ""; - my $statTopContributors = ""; - if( @topViews ) { - printMsg( " - top view: $topViews[0]", $cgiQuery ); - $statTopViews = join( "
", @topViews ); - } - if( @topContribs ) { - printMsg( " - top contributor: $topContribs[0]", $cgiQuery ); - $statTopContributors = join( "
", @topContribs ); - } - - # Update the WebStatistics topic - - my $tmp; - my $statsTopic = $TWiki::statisticsTopicname; - # DEBUG - # $statsTopic = "TestStatistics"; # Create this by hand - if( &TWiki::Store::topicExists( $webName, $statsTopic ) ) { - my( $meta, $text ) = &TWiki::Store::readTopic( $webName, $statsTopic, 1 ); - my @lines = split( /\n/, $text ); - my $statLine; - my $idxStat = -1; - my $idxTmpl = -1; - for( my $x = 0; $x < @lines; $x++ ) { - $tmp = $lines[$x]; - # Check for existing line for this month+year - if( $tmp =~ /$theLogMonthYear/ ) { - $idxStat = $x; - } elsif( $tmp =~ /<\!\-\-statDate\-\->/ ) { - $statLine = $tmp; - $idxTmpl = $x; - } - } - if( ! $statLine ) { - $statLine = "| | | | | | |"; - } - $statLine =~ s/<\!\-\-statDate\-\->/$theLogMonthYear/; - $statLine =~ s/<\!\-\-statViews\-\->/ $statViews/; - $statLine =~ s/<\!\-\-statSaves\-\->/ $statSaves/; - $statLine =~ s/<\!\-\-statUploads\-\->/ $statUploads/; - $statLine =~ s/<\!\-\-statTopViews\-\->/$statTopViews/; - $statLine =~ s/<\!\-\-statTopContributors\-\->/$statTopContributors/; - - if( $idxStat >= 0 ) { - # entry already exists, need to update - $lines[$idxStat] = $statLine; - - } elsif( $idxTmpl >= 0 ) { - # entry does not exist, add after line - $lines[$idxTmpl] = "$lines[$idxTmpl]\n$statLine"; - - } else { - # entry does not exist, add at the end - $lines[@lines] = $statLine; - } - $text = join( "\n", @lines ); - $text .= "\n"; - - &TWiki::Store::saveTopic( $webName, $statsTopic, $text, $meta, "", 1, 1, 1 ); - printMsg( " - Topic $statsTopic updated", $cgiQuery ); - - } else { - printMsg( " *** Warning: No updates done, topic $webName.$statsTopic does not exist", $cgiQuery ); - } -} - -# ========================= -# Get the items with top N frequency counts -# Items can be topics (for view hash) or users (for contrib hash) -sub getTopList -{ - my( $theMaxNum, $webName, $statsRef ) = @_; - - # Get reference to the sub-hash for this web - my $webhashref = $statsRef->{$webName}; - - # print "Main.WebHome views = " . $statsRef->{$webName}{'WebHome'}."\n"; - # print "Main web, TWikiGuest contribs = " . ${$statsRef}{$webName}{'Main.TWikiGuest'}."\n"; - - my @list = (); - my $topicName; - my $statValue; - - # Convert sub hash of item=>statsvalue pairs into an array, @list, - # of '$statValue $topicName', ready for sorting. - while( ( $topicName, $statValue ) = each( %$webhashref ) ) { - # Right-align statistic value for sorting - $statValue = sprintf "%7d", $statValue; - # Add new array item at end of array - $list[@list] = "$statValue $topicName"; - } - - # DEBUG - # print " top N list for $webName\n"; - # print join "\n", @list; - - # Sort @list by frequency and pick the top N entries - if( @list ) { - @list = map{ s/^([0-9][0-9][^0-9])/\ \;$1/; $_ } - map{ s/^([0-9][^0-9])/\ \;\ \;$1/; $_ } - map{ s/^\s*//; $_ } - reverse( sort( @list ) ); - if( $theMaxNum >= @list ) { - $theMaxNum = @list - 1; - } - return @list[0..$theMaxNum]; - } - return @list; -} - -# ========================= -sub printMsg -{ - my( $msg, $cgiQuery ) = @_; - my $htmlMsg = $msg; - - if( $cgiQuery ) { - # TODO: May need to fix this regex if internationalised script - # messages are supported in future. - if( $htmlMsg =~ /^[A-Z]/ ) { - $htmlMsg =~ s/^([A-Z].*)/

$1<\/h3>/go; - } else { - $htmlMsg =~ s/(\*\*\*.*)/$1<\/font>/go; - $htmlMsg =~ s/^\s\s/  /go; - $htmlMsg =~ s/^\s/ /go; - $htmlMsg .= "
"; - } - $htmlMsg =~ s/==([A-Z]*)==/==$1==<\/font>/go; - print "$htmlMsg\n"; - } else { - print "$msg\n"; - } -} - -# ========================= -# EOF diff --git a/tl b/tl deleted file mode 100755 index cf4f172..0000000 --- a/tl +++ /dev/null @@ -1,5 +0,0 @@ -#!/bin/sh - -limit=$1; shift - -task list limit:$limit "$@" diff --git a/tmutt b/tmutt deleted file mode 100755 index ba5b7ef..0000000 --- a/tmutt +++ /dev/null @@ -1,2 +0,0 @@ -#!/bin/sh -x-terminal-emulator -name mutt -e mutt "$@" diff --git a/to_eiche b/to_eiche deleted file mode 100644 index 83399fa..0000000 --- a/to_eiche +++ /dev/null @@ -1,20 +0,0 @@ - -# fstab - -cat /etc/fstab | sed 's,^/dev/discs/disc0/part3,#/dev/discs/disc0/part3,' > \ -/tmp/fstab -cat /tmp/fstab | sed 's,^/dev/discs/disc0/part1,#/dev/discs/disc0/part1,' > \ -/etc/fstab -cat /etc/fstab | sed 's,^\(/dev/root.*\)ext3,\1reiserfs,' > \ -/tmp/fstab -mv /tmp/fstab /etc/fstab - -# X - -cat /etc/X11/XF86Config | sed 's/\(Driver.*\)"ati"/\1"nv"/' > /tmp/xf86conf -cat /tmp/xf86conf | sed 's,\(Option.*"Protocol".*\)"ps/2",\1"MouseMan",' > \ -/etc/X11/XF86Config -cat /etc/X11/XF86Config | sed 's,\(Option.*"Device".*\)"/dev/misc/psaux",\1"/dev/tts/0",' > /tmp/xf86conf -mv /tmp/xf86conf /etc/X11/XF86Config - - diff --git a/to_flapp b/to_flapp deleted file mode 100644 index d39c99b..0000000 --- a/to_flapp +++ /dev/null @@ -1,20 +0,0 @@ - -# fstab - -cat /etc/fstab | sed 's,^#/dev/discs/disc0/part3,/dev/discs/disc0/part3,' > \ -/tmp/fstab -cat /tmp/fstab | sed 's,^#/dev/discs/disc0/part1,/dev/discs/disc0/part1,' > \ -/etc/fstab -cat /etc/fstab | sed 's,^\(/dev/root.*\)reiserfs,\1ext3,' > \ -/tmp/fstab -mv /tmp/fstab /etc/fstab - -# X - -cat /etc/X11/XF86Config | sed 's/\(Driver.*\)"nv"/\1"ati"/' > /tmp/xf86conf -cat /tmp/xf86conf | sed 's,\(Option.*"Protocol".*\)"MouseMan",\1"ps/2",' > \ -/etc/X11/XF86Config -cat /etc/X11/XF86Config | sed 's,\(Option.*"Device".*\)"/dev/tts/0",\1"/dev/misc/psaux",' > /tmp/xf86conf -mv /tmp/xf86conf /etc/X11/XF86Config - - diff --git a/torrent_handler/alle_screens.sh b/torrent_handler/alle_screens.sh deleted file mode 100755 index e8f2c0a..0000000 --- a/torrent_handler/alle_screens.sh +++ /dev/null @@ -1,6 +0,0 @@ -#!/bin/sh - -for screen in $(screen -ls | awk '/Detached/ { print $1 }'); do - echo screen -Ux $screen; - screen -Ux $screen; -done diff --git a/torrent_handler/downloader.sh b/torrent_handler/downloader.sh deleted file mode 100755 index 9e2cf20..0000000 --- a/torrent_handler/downloader.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh - -torrent="$1" -logfile="$2" -echo "Logging to $logfile" - -#btdownloadheadless --max_upload_rate 1 "$torrent" > "$logfile" 2>&1 -btdownloadheadless --display_interval 10 --max_upload_rate 1 "$torrent" 2>&1 | tee "$logfile" diff --git a/torrent_handler/find_succeeded.sh b/torrent_handler/find_succeeded.sh deleted file mode 100755 index 91cdffa..0000000 --- a/torrent_handler/find_succeeded.sh +++ /dev/null @@ -1,27 +0,0 @@ -#!/bin/sh -# Nico Schottelius -# find succussfully downloaded torrents - -TOPABS=$(cd $(dirname $0); pwd -P) -TORDIR="$TOPABS/torrents" -LOGDIR="$TOPABS/logs" -OUTDIR="$TOPABS/out" -SAVEDIR="$TOPABS/fertig" -SAVETOR="$SAVEDIR/torrents" -SEARCHTERM='time left: Download Succeeded!' -TESTLINES="20" - -# place downloads into current diretory/out -cd "$LOGDIR" - -#for torrent in ${TORDIR}/*; do -for log in *; do - tail -n${TESTLINES} "$log" | grep -q "$SEARCHTERM" >/dev/null - if [ "$?" -eq 0 ]; then - echo "$log" finished. - file=$(tail -n${TESTLINES} "$log" | awk -F: '/^download to:/ { print $2 }' | tail -n1 | sed 's;^ */;/;') - mv "${file}" "${SAVEDIR}" - mv "${TORDIR}/${log}" "${SAVETOR}" - rm "$log" - fi -done diff --git a/torrent_handler/searchterm b/torrent_handler/searchterm deleted file mode 100644 index ef89517..0000000 --- a/torrent_handler/searchterm +++ /dev/null @@ -1 +0,0 @@ -time left: Download Succeeded! diff --git a/torrent_handler/show_log_status.sh b/torrent_handler/show_log_status.sh deleted file mode 100755 index 830c843..0000000 --- a/torrent_handler/show_log_status.sh +++ /dev/null @@ -1,5 +0,0 @@ -for file in logs/*; - do - echo -n "${file}: "; - tail -n20 "$file" | grep "percent done:" | tail -n1; -done diff --git a/torrent_handler/start.sh b/torrent_handler/start.sh deleted file mode 100755 index af870bf..0000000 --- a/torrent_handler/start.sh +++ /dev/null @@ -1,22 +0,0 @@ -#!/bin/sh - -#set -x - - -TOPABS=$(cd $(dirname $0); pwd -P) -TORDIR="$TOPABS/torrents" -LOGDIR="$TOPABS/logs" -OUTDIR="$TOPABS/out" -DNLOADR="$TOPABS/downloader.sh" - - -rm -f "${LOGDIR}/"* -# place downloads into current diretory/out -cd "$OUTDIR" - -for torrent in ${TORDIR}/*; do - logfile="$LOGDIR/$(basename "$torrent")" - #echo screen -d -m btdownloadcurses --max_upload_rate 1 "$torrent" G "$logfile" - #screen -d -m btdownloadheadless --max_upload_rate 1 "$torrent" > "$logfile" - screen -d -m "$DNLOADR" "$torrent" "$logfile" -done diff --git a/torrent_handler/stop.sh b/torrent_handler/stop.sh deleted file mode 100755 index d5c9242..0000000 --- a/torrent_handler/stop.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh - -#PIDS=$(ps axuwww | awk '/\/usr\/bin\/btdownloadcurses/ { print $2 }') -PIDS=$(ps axuwww | awk '/\/usr\/bin\/btdownloadheadless/ { print $2 }') - -kill -TERM $PIDS -sleep 10 -kill -KILL $PIDS 2>/dev/null diff --git a/torrent_nach_hause b/torrent_nach_hause deleted file mode 100755 index f2299f4..0000000 --- a/torrent_nach_hause +++ /dev/null @@ -1,2 +0,0 @@ -#!/bin/sh -scp *torrent zuhause.schottelius.org:torrent/t/ && rm *torrent diff --git a/update-git+cogito.sh b/update-git+cogito.sh deleted file mode 100755 index a83e70a..0000000 --- a/update-git+cogito.sh +++ /dev/null @@ -1,86 +0,0 @@ -#!/bin/sh -# Nico Schottelius -# Date: Mit Nov 9 12:57:04 CET 2005 -# Last Modified: Thu Nov 10 12:53:00 CET 2005 -# -# Install latest git and cogito to $INSTALL_PREFIX/$name and -# link the binaries to $BINDIR -# Needs git and cogito to work! -# And: -# - libexpat-dev -# - libcurl3-dev -# - libssl-dev -# - libz-dev - - -# -# Configuration -# -BUILDDIR=${BUILDDIR:-/home/user/nico/build} -INSTALL_PREFIX=${INSTALL_PREFIX:-/usr/packages} -BINDIR=${BINDIR:-/usr/local/bin} -MAKE=${MAKE:-make} - - -# -# No configuration needed below here -# - -BASE_GET=http://www.kernel.org/pub/scm/ -PROJECTS="git cogito" -# urls for initial cloning -git_url=http://www.kernel.org/pub/software/scm/git/ -cogito_url=http://www.kernel.org/pub/software/scm/cogito/ - -error_in() -{ - echo "[failed] $1" - echo "Exiting now." - exit 1 -} - - -for project in $PROJECTS; do - echo "Working on $project (in $BUILDDIR/$project) ... " - - - # Test for existence: if git or cogito is missing, install it from - # the current archive - which $project > /dev/null - - if [ $? -ne 0 ]; then - echo "$project not found, installing from archive" - - eval base_url=\$${project}_url - get_url=$(lynx -dump "$base_url" | awk "/$project-([0-9]\.)*tar.bz2\$/ { print \$2 }" | sort -n | tail -n 1) - - echo "Retrieving $project from $get_url" - fi - - if [ ! -d "$BUILDDIR/$project" ]; then - echo "Cloning $project" - cg-clone "${BASE_GET}${project}/${project}.git" "$BUILDDIR/$project" - else - echo "Updating $project from \"origin\"" - cd "$BUILDDIR/$project" || error_in "$BUILDDIR/$project" - cg-update origin - fi - - if [ $? -ne 0 ]; then - echo "Pull or clone failed, ABORTING." - exit 23 - fi - - cd "$BUILDDIR/$project" - version=$(cg-object-id) - echo "Installing $project (Version: $version)" - DDIR=$INSTALL_PREFIX/$project-$version - $MAKE clean || error_in "Cleaning $project" - $MAKE "prefix=$DDIR" all || error_in "Building $project" - $MAKE "prefix=$DDIR" install || error_in "Installing $project" - - echo "Linking files to $BINDIR ..." - for file in "$DDIR/bin/"*; do - ln -sf $file "$BINDIR" || error_in "Linking $file" - done -done diff --git a/update-puppet-on-host b/update-puppet-on-host deleted file mode 120000 index bbba05a..0000000 --- a/update-puppet-on-host +++ /dev/null @@ -1 +0,0 @@ -/home/users/nico/localch/vcs/sysadmin-logs/update-puppet-on-host \ No newline at end of file diff --git a/vim-term b/vim-term deleted file mode 100755 index f375aca..0000000 --- a/vim-term +++ /dev/null @@ -1,5 +0,0 @@ -#!/bin/sh -# Open vim - -# x-terminal-emulator -e vim "$@" -urxvt -e vim "$@" diff --git a/wlan b/wlan deleted file mode 100755 index ffdcb8c..0000000 --- a/wlan +++ /dev/null @@ -1,50 +0,0 @@ -#!/bin/sh -e -# Nico Schottelius, 20080910 11:09 -# The initial version -# -# udhcpc -nfqi wlp0s3 && vpnc ~nico/ethz/vpn/pc.conf -# -# Failed due to many reasons. See the working hack/version below. -# - -if [ "$#" -ge 1 ]; then - dev=$1; shift -else - dev=wlp3s0 - dev=wlan0 - dev=wlp2s0 - dev=wlp1s0 - dev=wlp4s0 - dev=wlp3s0 -fi - -set -x -set +e -wpa_cli terminate || true -#rmmod b43 bcma -sleep 1 -#modprobe b43 -set -e - -#ip l s eth0 down -# Ensure previous wpa_supplicant is not running -( wpa_cli terminate || exit 0 ) - -# Ensure the adapter works, reloading the driver helps -# ( rmmod iwlagn; modprobe iwlagn; sleep 2 ) - -# wpa_supplicant -B -Dwext -iwlp0s3 -c /etc/wpa_supplicant/wpa_supplicant.conf -# wpa_supplicant -B -Dwext -i$dev -c /etc/wpa_supplicant/wpa_supplicant.conf -#wpa_supplicant -B -Dnl80211 -i$dev -c /etc/wpa_supplicant/wpa_supplicant.conf -wpa_supplicant -B -Dnl80211 -i$dev -c /etc/wpa_supplicant/wpa_supplicant-${dev}.conf -# Give wpa some time to connect -sleep 5 - -# Choose an dhcp client -( udhcpc -nqfi $dev || dhcpcd ) - -wpa_cli - -# vpnc ~nico/ethz/vpn/sg.conf -#vpnc ~nico/firmen/ethz/vpn/vpnc.conf -# vpnc-disconnect diff --git a/x-links b/x-links deleted file mode 100755 index c6dd090..0000000 --- a/x-links +++ /dev/null @@ -1,5 +0,0 @@ -#!/bin/sh -x-off - -xrandr --output HDMI1 --auto --rotate normal --left-of eDP1 -xrandr --output DP1 --auto --rotate normal --left-of eDP1