CVC3 is an automatic theorem prover for Satisfiability Modulo Theories (SMT)
problems. It can be used to prove the validity (or, dually, the
satisfiability) of first-order formulas in a large number of built-in logical
theories and their combination.
CVC3 is the last offspring of a series of popular SMT provers, which originated
at Stanford University with the SVC system. In particular, it builds on the
code base of CVC Lite, its most recent predecessor. Its high level design
follows that of the Sammy prover.
CVC3 works with a version of first-order logic with polymorphic types and has
a wide variety of features including:
* several built-in base theories: rational and integer linear arithmetic,
arrays, tuples, records, inductive data types, bit vectors, and equality
over uninterpreted function symbols;
* support for quantifiers;
* an interactive text-based interface;
* a rich C and C++ API for embedding in other systems;
* proof and model generation abilities;
* predicate subtyping;
* essentially no limit on its use for research or commercial purposes
(see license).
Book of Psalms from the Douai Bible (1610) in fortune(6) file format
In general, taking random out-of-context verses from a Bible is a
VERY bad idea. This said, this is an experiment to generate a not
too bad fortune-cookie database under the following principles:
- Only the text from the psalms, which are usually meant for praying,
were taken.
- The texts always include at least some context: you are always
notified where the text came from and there is always sufficient
text so that the phrases make sense.
- The text was taken from the classic Douai Bible, a direct translation
from the latin Vulgata. The old language will not give you false
impressions that you are actually understanding it fully.
- No effort was done to remove the original comments. Non-Christians
may rightfully feel the translation is biased.
This is meant to be a general aid for Christian meditation: it is not
generally to be taken as my message-of-the-day from God.
To use, you need UNIX fortune(6) utility and you should follow the
instructions from the corresponding man page.
[ excerpt from developer's site ]
It is a free library for decoding mpeg-2 and mpeg-1 video
streams. The main goals in libmpeg2 development are:
Conformance - libmpeg2 is able to decode all mpeg streams that
conform to certain restrictions: "constrained parameters" for
mpeg-1, and "main profile" for mpeg-2. In practice, this is what
most people are using. For streams that follow these restrictions,
we believe libmpeg2 is 100% conformant to the mpeg standards - and
we have a pretty extensive test suite to check this.
Speed - for most current systems, the display will actually take
more time than the mpeg-2 decoding. For systems that have hardware
color conversion and scaling (as we can use with the xv extension
in Xfree 4), you should be able to watch DVD streams on a Celeron
400. On a PIII/666 with null display you should get about 110 frames
per second.
Portability - most of the code is written in C, and when we use
platform-specific optimizations we always have a generic C routine
to fall back on.
Radiator is a highly configurable and flexible Radius server that supports
authentication by a huge range of authentication methods such as Flat files,
DBM files, Unix password files, SQL databases, remote Radius servers
(proxying), external programs, NT User Manager, Active Directory, LDAP, PAM,
iPASS, GRIC, NIS+, Tacacs+, a wide range of ISP billing packages such as
Emerald, Platypus, Rodopi, Hawk-i, Interbiller98, Freeside etc, your legacy
user database etc, etc.
Radiator now supports more 802.1x secure wireless and LAN authentication
methods than any other Radius server giving a wide choice of 802.1x network
clients.
Radiator also includes many features not found in other Radius servers such
as double-login prevention, username rewriting, full vendor-specific
attributes, time-of-day blocking and a GUI for running user tests. Full list
of technical features.
Runs on all Unix, Linux, Windows 95, 98, NT, 2000, Mac OS-9 and Mac OS-X, VMS.
Due to license restrictions, this package must be purchased and manually
downloaded from the Open System Consultants web site.
If you have an AT&T Wireless, Bell Canada/Bell Mobility, Cellular One,
Cingular, Cricket, Sprint PCS, SkyTel, or T-Mobile cell phone or pager, and you
want the ability to send SMS messages to it via a command-line utility, this is
what you need. All this program requires is a computer with a baseline Perl 5.x
installation and web access. NO EXTRA PERL MODULES REQUIRED!
How does it work?
SendSMS connects to your service provider's web page and pretends to submit a
form to their 'Instant Messaging' web page. Currently, AT&T Wireless, Bell
Canada/Bell Mobility, Cellular One, Cingular, Cricket, SkyTel, Sprint PCS, and
T-Mobile are supported. Users are encouraged to modify the provided templates to
add support for any providers who are currently unsupported.
Other Service Providers
If you are interested in supporting another service provider please try to
modify sendSMS on your own. It is not hard at all. Instructions and examples are
included in the code, and if you're familiar with the site you're porting to, it
takes about 15 minutes. If you get sendSMS working with any other providers' web
sites, please email Paul Kreiner [deacon at thedeacon.org] and/or the port
maintainer a patch so it can be added to the next release.
Preppy is ReportLab's templating system. It was developed in late 2000 and has
been in continual production use since then. It is open source (BSD-license).
The key features are:
- *small*. Preppy is a single Python module. If you want a templating system
'in the box', it's easy to include it in your project
- *easy to learn*. It takes about one minute to scan all the features
- *just Python*. We have not invented another language, and if you want to do
something - includes, quoting, filters - you just use Python
- *compiled to bytecode*: a .prep file gets compiled to a Python function in
a .pyc file
- *easy to debug*: preppy generates proper Python exceptions, with the correct
line numbers for the .prep file. You can follow tracebacks from Python
script to Preppy template and back, through multiple includes
- *easy to type and read*. We've been using ``{{this}}`` syntax since well
before Django was thought of
- *8-bit safe*: it makes no assumption that you are generating markup and does
nothing unexpected with whitespace; you could use it to generate images or
binary files if you wanted to.
DenyHosts is a script intended to be run by *ix system administrators to
help thwart ssh server attacks.
If you've ever looked at your ssh log (/var/log/auth.log ) you may be alarmed
to see how many hackers attempted to gain access to your server.
Denyhosts helps you:
- Parses /var/log/auth.log to find all login attempts
- Can be run from the command line, cron or as a daemon (new in 0.9)
- Records all failed login attempts for the user and offending host
- For each host that exceeds a threshold count, records the evil host
- Keeps track of each non-existent user (eg. sdada) when a login attempt failed.
- Keeps track of each existing user (eg. root) when a login attempt failed.
- Keeps track of each offending host (hosts can be purged )
- Keeps track of suspicious logins
- Keeps track of the file offset, so that you can reparse the same file
- When the log file is rotated, the script will detect it
- Appends /etc/hosts.allow
- Optionally sends an email of newly banned hosts and suspicious logins.
- Resolves IP addresses to hostnames, if you want
The ssh library was designed to be used by programmers needing a working SSH
implementation by the mean of a library. The complete control of the client is
made by the programmer. With libssh, you can remotely execute programs, transfer
files, use a secure and transparent tunnel for your remote programs.
With its Secure FTP implementation, you can play with remote files easily,
without third-party programs others than libcrypto (from openssl).
libssh features:
* Full C library functions for manipulating a client-side SSH connection
* SSH2 and SSH1 protocol compliant
* Fully configurable sessions
* Server support, SSH agent authentication support
* Support for AES-128, AES-192, AES-256, Blowfish, 3DES in CBC mode
* Use multiple SSH connections in a same process, at same time
* Use multiple channels in the same connection
* Thread safety when using different sessions at same time
* POSIX-like SFTP implementation with openssh extension support
* SCP implementation
* RSA and DSS server public key supported
* Compression support (with zlib)
* Public key (RSA and DSS), password and keyboard-interactive authentication
Bytes::Random::Secure provides two interfaces for obtaining crypto-quality
random bytes. The simple interface is built around plain functions. For greater
control over the Random Number Generator's seeding, there is an Object Oriented
interface that provides much more flexibility.
The "functions" interface provides functions that can be used any time you need
a string of a specific number of random bytes. The random bytes are available as
simple strings, or as hex-digits, Quoted Printable, or MIME Base64. There are
equivalent methods available from the OO interface, plus a few others.
This module can be a drop-in replacement for Bytes::Random, with the primary
enhancement of using a cryptographic-quality random number generator to create
the random data. The random_bytes function emulates the user interface of
Bytes::Random's function by the same name. But with Bytes::Random::Secure the
random number generator comes from Math::Random::ISAAC, and is suitable for
cryptographic purposes. The harder problem to solve is how to seed the
generator. This module uses Crypt::Random::Seed to generate the initial seeds
for Math::Random::ISAAC.
Unix provides the standard du utility, which scans your disk and tells you which
directories contain the largest amounts of data. That can help you narrow your
search to the things most worth deleting.
However, that only tells you what's big. What you really want to know is what's
too big. By itself, du won't let you distinguish between data that's big because
you're doing something that needs it to be big, and data that's big because you
unpacked it once and forgot about it.
Most Unix file systems, in their default mode, helpfully record when a file was
last accessed. Not just when it was written or modified, but when it was even
read. So if you generated a large amount of data years ago, forgot to clean it
up, and have never used it since, then it ought in principle to be possible to
use those last-access time stamps to tell the difference between that and a
large amount of data you're still using regularly.
agedu is a program which does this. It does basically the same sort of disk scan
as du, but it also records the last-access times of everything it scans. Then it
builds an index that lets it efficiently generate reports giving a summary of
the results for each subdirectory, and then it produces those reports on demand.