htop is an enhanced version of top, the interactive process viewer,
which can display the list of processes in a tree form.
Comparison between 'htop' and 'top'
* In 'htop' you can scroll the list vertically and horizontally
to see all processes and full command lines.
* In 'top' you are subject to a delay for each unassigned
key you press (especially annoying when multi-key escape
sequences are triggered by accident).
* 'htop' starts faster ('top' seems to collect data for a while
before displaying anything).
* In 'htop' you don't need to type the process number to
kill a process, in 'top' you do.
* In 'htop' you don't need to type the process number or
the priority value to renice a process, in 'top' you do.
* In 'htop' you can kill multiple processes at once.
* 'top' is older, hence, more tested.
chyves is a bhyve front-end manager. chyves manages type-2 virtualized guests by
utilizing hardware virtualization on a base FreeBSD 10.3+ installation. On a
base install, only FreeBSD guests can run. However, with the installation of
sysutils/grub2-bhyve and sysutils/bhyve-firmware from ports or pkg, most other
OSes can run as a guest, including Windows. See DEPENDENCIES section in the man
page for more information.
chyves is targeted for beginners as well as power users. Beginners should find
chyves relatively easy to use with lots of documentation and demonstrations.
While power users should find utility with features such as true ZFS clones,
PCI passthrough, rapid execution against many guests, disk images, and snapshot
reverted states on boot/reboot to name a few of the advanced features.
The name 'chyves' is the pluralized, big endian alphabetic increment of bhyve.
'chyves' is pronounced like 'chives', part of the Allium genus. The onion is
also in the Allium genus.
The LibXDiff library implements basic and yet complete functionalities to
create file differences/patches to both binary and text files. The library
uses memory files as file abstraction to achieve both performance and
portability. For binary files, LibXDiff implements (with some modification)
the algorithm described in File System Support for Delta Compression by
Joshua P. MacDonald, while for text files it follows directives described in
An O(ND) Difference Algorithm and Its Variations by Eugene W. Myers. Memory
files used by the library are basically a collection of buffers that store the
file content. There are two different requirements for memory files when passed
to diff/patch functions. Text files for diff/patch functions require that a
single line do not have to spawn across two different memory file blocks.
Binary diff/patch functions require memory files to be compact. A compact
memory files is a file whose content is stored inside a single block.
This is a smaller, cheaper, faster SED implementation. Minix uses it. GNU
used to use it, until they built their own sed around an extended (some
would say over-extended) regexp package.
For embedded use we searched for a tiny sed implementation especially for
use with the dietlibc and found Eric S. Raymond's sed implementation quite
handy. Though it suffered several bugs and was not under active maintenance
anymore. After sending a bunch of fixes we agreed to continue maintaining
this lovely, historic sed implementation.
Along a lot fixes and cleanups, further speedups, and some missing features
and POSIX conformance, we also added a test-suite to the package, so
regressions are quickly and easily uncovered.
String::ToIdentifier::EN provides a utility method, "to_identifier" for
converting an arbitrary string into a readable representation using the ASCII
subset of \w for use as an identifier in a computer program. The intent is to
make unique identifier names from which the content of the original string can
be easily inferred by a human just by reading the identifier.
If you need the full set of \w including Unicode, see the subclass
String::ToIdentifier::EN::Unicode.
Currently, this process is one way only, and will likely remain this way.
The default is to create camelCase identifiers, or you may pass in a separator
char of your choice such as _.
Binary char groups will be separated by _ even in camelCase identifiers to make
them easier to read, e.g.: foo_2_0xFF_Bar.
This program has three modes of operation:
- First, is ** EXPLOSION **, or the expanding of a .LIT file into an
OEBPS compliant package.
- Second, is the DOWNCONVERTING of a .LIT file down to "Sealed",
or DRM1 format for reading on handheld devices.
- Third, is the INSCRIBING of a .LIT file which allows you to label
your ebooks.
DRM5 is supported if you have a "keys.txt" file that contains
the private key(s) for your passport(s) in either the CLIT program
directory or the current directory.
This is a tool for **YOUR OWN FAIR USE** and not for stealing
other people's ebooks.
Please do not use this program to distrbute illegal copies of ebooks.
... that would make Baby Jesus sad.
This module provides an XPath engine, that can be re-used by other
module/classes that implement trees.
In order to use the XPath engine, nodes in the user module need to
mimick DOM nodes. The degree of similitude between the user tree and a
DOM dictates how much of the XPath features can be used. A module
implementing all of the DOM should be able to use this module very
easily (you might need to add the cmp method on nodes in order to get
ordered result sets).
This code is a more or less direct copy of the XML::XPath module by
Matt Sergeant. I only removed the XML processing part to remove the
dependency on XML::Parser, applied a couple of patches, renamed a
whole lot of methods to make Pod::Coverage happy, and changed the docs.
NLTK is a leading platform for building Python programs to work with human
language data. It provides easy-to-use interfaces to over 50 corpora and
lexical resources such as WordNet, along with a suite of text processing
libraries for classification, tokenization, stemming, tagging, parsing,
and semantic reasoning, and an active discussion forum.
Thanks to a hands-on guide introducing programming fundamentals alongside
topics in computational linguistics, NLTK is suitable for linguists,
engineers, students, educators, researchers, and industry users alike.
NLTK is available for Windows, Mac OS X, and Linux. Best of all, NLTK is
a free, open source, community-driven project.
NLTK has been called "a wonderful tool for teaching, and working in,
computational linguistics using Python" and "an amazing library to play
with natural language".
TECkit (Text Encoding Conversion toolkit) is a toolkit for converting data
between 8-bit legacy encodings and Unicode. It can also be used for
transliteration of Unicode between different scripts.
TECkit uses a mapping description language (mapping byte encodings to Unicode).
Mapping rules can be extended by (1) the use of character sequences rather than
single characters on either side; (2) by the addition of contextual constraints
(environments) determining when a rule should apply; (3) and by the use of
character classes, optional and repeatable elements, grouping and alternation
to express more complex patterns to be matched and processed.
TECkit is particularly useful with XeTeX (Unicode-aware derivate of TeX).
The following binaries are provided:
teckit_compile mapping compiler that allows binary mapping tables (.tec)
to be built from TECkit description files (.map)
sfconv a tool for converting Standard Format (SF) files
txtconv a utility to apply TECkit mappings to plain-text files
http://scripts.sil.org/TECkitDownloads#5b6cf869
__ ___ __ _
\ \ / / \/ | | Website META Language
\ \ /\ / / |\/| | |
\ V V /| | | | |___ ``WML is the Unix toolkit for getting
\_/\_/ |_| |_|_____| your webdesigner's HTML job done.''
Copyright (c) 1996-2002 Ralf S. Engelschall
Copyright (c) 1999-2002 Denis Barbier
WML is a free and extensible Webdesigner's off-line HTML generation toolkit
for Unix, distributed under the GNU General Public License (GPL v2). It is
written in ANSI C and Perl 5, build via a GNU Autoconf based source tree
and runs out-of-the-box on all major Unix derivates. It can be used
free of charge both in educational and commercial environments.
WML consists of a control frontend driving up to nine backends in a
sequential pass-oriented filtering scheme. Each backend provides one
particular core language. For maximum power WML additionally ships with a
well-suited set of include files which provide higher-level features build
on top of the backends core languages. While not trivial and idiot proof WML
provides most of the core features real hackers always wanted for HTML
generation.