mod_geoip2 is an Apache2 module that looks up the country code for the
IP address making the request without using reverse DNS.
This is a very small (7 Kb) HTTP server. It runs from inetd, which means
its performance is poor. But for low-traffic sites, it's quite adequate.
It implements all the basic features of an HTTP server.
The crawl utility starts a depth-first traversal of the web at the
specified URLs. It stores all JPEG images that match the configured
constraints. Crawl is fairly fast and allows for graceful termination.
After terminating crawl, it is possible to restart it at exactly
the same spot where it was terminated. Crawl keeps a persistent
database that allows multiple crawls without revisiting sites.
The main reason for writing crawl was the lack of simple open source
web crawlers. Crawl is only a few thousand lines of code and fairly
easy to debug and customize.
Some of the main features:
- Saves encountered JPEG images
- Image selection based on regular expressions and size contrainsts
- Resume previous crawl after graceful termination
- Persistent database of visited URLs
- Very small and efficient code
- Supports robots.txt
This add-on module allows the apache web server to use a PostgreSQL
database for user and/or group authentication. For large user lists this
can offer a significate speed up over apache's standard flat file
format.
mod_extract_forwarded hooks itself into Apache's header parsing phase and looks
for the X-Forwarded-For header which some (most?) proxies add to the proxied
HTTP requests. It extracts the IP from the X-Forwarded-For and modifies the
connection data so to the rest of Apache the request looks like it came from
that IP rather than the proxy IP.
mod_extract_forwarded can be dangerous for host based access control because
X-Forwarded-For is easily spoofed. Because of this you can configure which
proxies you trust or don't trust.
mod_fileiri implements http IRIs for directories/files, i.e.
if accepts URIs with non-ASCII characters encoded in UTF-8 and
converts them to the legacy encoding used in the file system
(which can be specified per directory, or even finer if necessary
(although that's a real hack)).
What is more, it continues to accept requests in the legacy
encoding specified, and redirects them to the correct UTF-8
form, which then returns the actual document (without looping).
There is also a backwards mode, which does redirects from
URIs in a specified legacy encoding to UTF-8 if the directory/
filenames are in UTF-8.
mod_hosts_access allows you to use the hosts.allow and hosts.deny
files to configure access to your Apache webserver. This is the
stable version 1.0.0. Nothing has changed apart from the homepage
and some documentation glitches.
Fast templating system using the same syntax as HTML::Template but
which compiles templates to perl code.
Language packs for Firefox
Log Validator is a web server log analysis tool with focus on the quality
of Web documents. Thanks to a modular, extensible design, the Log Validator
can help Web authors find the most popular content on their web site that
matches a particular criteria.
The Log Validator was first written with Validation (HTML, etc.) in mind :
it can thus help web content managers find and fix the most frequently
accessed invalid documents on their Web site, acting as a comprehensive,
step-by-step validation tool.