Dblatex started as a DB2LaTeX clone. So, why this project? The purpose
is a bit different on these points:
(1) The project is end-user oriented, that is, it tries to hide as much
as possible the latex compiling stuff by providing a single clean
script to produce directly DVI, PostScript and PDF output.
(2) The actual output rendering is done not only by the XSL stylesheets
transformation, but also by a dedicated LaTeX package. The purpose is
to allow a deep LaTeX customisation without changing the XSL
stylesheets.
(3) Post-processing is done by Python, to make publication faster,
convert the images if needed, and do the whole compilation.
XML::Node is a Perl5 module which provides a simplified extension interface
to XML::Parser.
Paraphrasing the README:
Instead of worrying about "start", "end", and "char" callbacks of every
single XML node, you can simply say that you only want to be notified when
a path is found.
Using XML::Node, you can ignore the parts of XML files that you are not
interested in. Additionally, you can register a variable instead of a
callback function. The corresponding string found in an XML file will be
automatically appended to your variable.
RXP is a very fast validating XML parser written by Richard Tobin
of the University of Edinburgh. It complies fully with the W3C test
suites (although we have compiled it without Unicode support for
the time being). pyRXP is a wrapper around this which constructs a
lightweight in-memory "tuple tree" in a single call. This structure
is the lightest one we could define in Python, and it is constructed
entirely in C code, resulting in unprecedented speed. It is a core
part of ReportLab's forthcoming XML toolkit, which aims to offer
simple, fast and pythonic tools for common XML processing tasks.
TagSoup - Just Keep On Truckin'
TagSoup is a SAX-compliant parser written in Java that, instead of parsing
well-formed or valid XML, parses HTML as it is found in the wild: poor,
nasty and brutish, though quite often far from short. TagSoup is designed
for people who have to process this stuff using some semblance of a rational
application design. By providing a SAX interface, it allows standard XML
tools to be applied to even the worst HTML. TagSoup also includes
a command-line processor that reads HTML files and can generate either
clean HTML or well-formed XML that is a close approximation to XHTML.
XML/Ada is a full XML suite for use with Ada compilers, such as GNAT AUX.
XML/Ada is a set of modules that provide a simple manipulation of XML
streams. It supports the whole XML 1.1 specification and can parse any file
that follows this standard, including the contents of the DTD although no
validation of the documents is performed based on those.
It provides support for a number of standards associated with XML such as
SAX, DOM, and XML schemas. Additionally, it includes a module to manipulate
unicode streams since this is required by the XML standard.
Apache Cocoon is a web development framework built around the
concepts of separation of concerns and component-based web development.
Cocoon implements these concepts around the notion of 'component
pipelines', each component on the pipeline specializing on a
particular operation. This makes it possible to use a Lego(tm)-like
approach in building web solutions, hooking together components
into pipelines without any required programming.
Cocoon is "web glue for your web application development needs".
It is a glue that keeps concerns separate and allows parallel
evolution of all aspects of a web application, improving development
pace and reducing the chance of conflicts.
Hugo is a general-purpose website framework. Technically speaking, Hugo is a
static site generator. Unlike other systems which dynamically build a page
every time a visitor requests one, Hugo does the building when you create your
content. Since websites are viewed far more often than they are edited, Hugo is
optimized for website viewing while providing a great writing experience.
Sites built with Hugo are extremely fast and very secure. Hugo sites can be
hosted anywhere and run without dependencies on expensive runtimes like Ruby,
Python or PHP and without dependencies on any databases.
Beautiful Soup parses arbitrarily invalid XML- or HTML-like substance
into a tree representation. It provides methods and Pythonic idioms
that make it easy to search and modify the tree.
A well-formed XML/HTML document will yield a well-formed data
structure. An ill-formed XML/HTML document will yield a
correspondingly ill-formed data structure. If your document is only
locally well-formed, you can use this library to find and process the
well-formed part of it. The BeautifulSoup class has heuristics for
obtaining a sensible parse tree in the face of common HTML errors.
This module takes a hierarchy of directories containing MP3 files
and presents it as a browsable song library for streaming over the
web. It requires the Apache web server, the mod_perl embedded Perl
interpreter, and the MP3::Info module.
MP3 files are displayed in a list that shows the MP3 title, artist,
duration and bitrate. Subdirectories are displayed with "CD" icons.
The user can download an MP3 file to disk by clicking on its title,
stream it to an MP3 decoder by clicking on the "play" link. Users
can also stream the entire contents of a directory, or select a
subset of songs to play.
This module provide a simple interface to filter entries out
of an httpd logfile. The constructor can be passed regular
expressions to match against particular fields on the
logfile. It does its filtering line by line, using a filter
method that takes a line of a logfile as input, and returns
true if it matches, and false if it doesn't.
There are two possible non-matching (false) conditions; one
is where the line is a valid httpd logfile entry, but just
doesn't happen to match the filter (where "" is returned).
The other is where it is an invalid entry according to the
format specified in the constructor.