Tags : Browse Projects

Select a tag to browse associated projects and drill deeper into the tag cloud.

gama-gateway

Compare

  Analyzed about 10 hours ago

Gama Gateway RDF Repository and GAMA data model

22.9K lines of code

0 current contributors

almost 7 years since last commit

0 users on Open Hub

Inactive
0.0
 
I Use This

Archive::Zip Perl Module

Compare

  Analyzed about 22 hours ago

The Archive::Zip module allows a Perl program to create, manipulate, read, and write Zip archive files. Zip archives can be created, or you can read from existing zip files. Once created, they can be written to files, streams, or strings. Members can be added, removed ... [More] , extracted, replaced, rearranged, and enumerated. They can also be renamed or have their dates, comments, or other attributes queried or modified. Their data can be compressed or uncompressed as needed. Members can be created from members in existing Zip files, or from existing directories, files, or strings. This module uses the Compress::Zlib library to read and write the compressed streams inside the files. One can use Archive::Zip::MemberRead to read the zip file archiv [Less]

6.11K lines of code

8 current contributors

over 1 year since last commit

0 users on Open Hub

Very Low Activity
0.0
 
I Use This

chat-archive

Compare

  Analyzed 1 day ago

Easy to use offline chat archive

2.51K lines of code

0 current contributors

about 4 years since last commit

0 users on Open Hub

Inactive
0.0
 
I Use This

WARCreate

Compare

  Analyzed about 1 hour ago

Google Chrome extension to "Create WARC files from any webpage".

1.87K lines of code

2 current contributors

5 months since last commit

0 users on Open Hub

Very Low Activity
0.0
 
I Use This

archivefs

Compare

  Analyzed about 1 hour ago

AboutArchiveFS is a FUSE file system used for archiving and backup. Its primary function is to ensure that multiple copies of a file are only represented as a single file. The representation of the file system is intentionally kept simple and consists just of a single SQLite3 database file and ... [More] table (which can be dumped into a text file), together with a directory full of files. The file system is not intended for general purpose computing, but mostly for copying data in and out. It seems to be working reasonably well for backup, and even file system intensive operations like software builds seem to complete OK. Please give it a good try and workout, but don't blame me if you lose any data. UsageJust check out the source code. You do need the python-fuse and python-sqlite3 packages (Ubuntu) or their equivalents. To start it up, use a command like: $ python archivefs.py -o root=/somewhere/FSDATA /my/mountpoint $ echo hello world > /my/mountpoint/new-file $ cat /my/mountpoint/new-fileThe root directory must exist and be writable by you. The root directory contains the database file (DB), a working directory for temporary files (WORKING), and an archival directory containing the actual, permanent files (ARCHIVE). The file system will create those if they don't already exist. When you're done, you should unmount the directory as usual: $ fusermount -u /my/mountpointIt's intended to be used with something like: cp -av /home/tmb /backup/tmb-$(date)You can get some file metadata via getfattr and attr: attr -g _id file -- the unique file id attr -g _storage file -- the path to the actual file attr -g _instances file -- a list of all paths referring to this content Note the following points: file permissions aren't enforced (but are recorded) link counts are not preserved deleting a file only deletes its entry, it doesn't recover the space automatically There are a number of things I can't find good documentation and that I therefore don't quite understand in fuse-python: hardlinks and concurrent updates through different paths the degree of threading (apparently, not much, but enough to cause occasional problems) how mmap is handled You can reconstruct a directory tree easily from an md5sum dump and the contents of the archive disk; you don't need FUSE. To create such a dump manually, just write: $ find . -type f -print0 | xargs -0 md5sum > my.md5sums(I'll upload some scripts for this at some point.) HistoryThis code replaces (and is based on) a bunch of shell scripts I've been using for backup for a couple of decades that also used checksums for storage but stored the mapping in a plain text file. The reason why a file system is nicer than the scripts is because it's possible not only to copy into the archival tree, but also untar tar files in it directly, copy data in remotely, etc. With FUSE, it's finally easy and portable enough to do this (last time I looked into doing this, this still required a lot of painful kernel-level C programming.) InternalsIt's written in Python using the python-fuse package. The representation of the file system is pretty simple: root/DB -- sqlite3 database file containing metadata and ids root/ARCHIVE/xx/yy/xxyyzzzzz... -- the actual content, stored by id to keep directory size down, this has two levels of directories root/WORKING/zzzzzzzz... -- temporary working files TODOThere are a bunch of things to be done: important clean up the code write a text file dumper for the database smart command line tools for local and remote copies/sync garbage collecting defunct working files on startup garbage collecting defunct archival files on demand (after a big removal) automatic garbage collection of defunct archival files upon deletion add metadata handling and searchrecord checksum and discard) well-known checksums (just transparent gzip compression/decompression of chunks would be nice record-and-discard well-known checksum (can retrieve from the web, maybe store URL) by file name by mime type separate directory and file name columns to make dir listings faster tokenize directory names to save space id available via extended attribute speed it up by caching and other tricks better multithreading (maybe port to IronPython) record user ids in text form and resolve at runtime fix global scope for fs variable transparently handle files inside archives write a test suite and perform more extensive testing perform explicit in-memory buffering for checksumming and copying use a larger checksum to make collisions less likely add non-FUSE command line tools for storing and accessing the data handle extended attributes tools for reporting logical vs physical usage move small file operations in memory transparent mounting of the underlying file system long term ideas (maybe a different project) handle file parts by partitioning files at type-dependent boundaries e.g., paragraph boundaries, MP3 chunks, mbox message boundaries, etc. transparently disassemble and assemble archive formats S3 backend stick very small files into the database distributed storage across disks distributed storage across the network change tracking time-machine like functionality i.e. represent trees at different points in time explicitly also saves database space for frequent backups this needs to have a notion of a completed checkpoint, so... archivefs-open-replica old-tree new-tree rsync ... source new-tree archivefs-close-replica new-tree old-tree [Less]

486 lines of code

0 current contributors

over 9 years since last commit

0 users on Open Hub

Inactive
0.0
 
I Use This

datadiver

Compare

  Analyzed 1 day ago

A library and application for data analysis and extraction.

676 lines of code

0 current contributors

almost 10 years since last commit

0 users on Open Hub

Inactive
0.0
 
I Use This

comodojo/zip

Compare

  Analyzed about 13 hours ago

ZipArchive toolbox

1.23K lines of code

2 current contributors

over 2 years since last commit

0 users on Open Hub

Inactive
0.0
 
I Use This

gnochive

Compare

  No analysis available

gnochive is a GUI frontend for all common archivers under linux. gnochive was formally known as gnomerar Supports: - rar - gzip - tar - zip/unzip

0 lines of code

0 current contributors

0 since last commit

0 users on Open Hub

Activity Not Available
0.0
 
I Use This
Mostly written in language not available
Licenses: gpl

WP Publication Archive

Compare

  Analyzed about 11 hours ago

Based on the abandoned wp-publications-archive project hosted on Google Code! WP Publication Archive adds a custom content type for storing, tagging, and categorizing downloadable content external to standard WordPress posts and pages. You can add downloadable PDF files, Word documents, and ... [More] PowerPoint presentations. These files will be stored in the standard WordPress uploads directory but will be managed separately through a custom post type interface in the WordPress admin area. [Less]

3.52K lines of code

0 current contributors

over 9 years since last commit

0 users on Open Hub

Inactive
5.0
 
I Use This

distribute-on{to,}-CDs

Compare

  Analyzed about 12 hours ago

Distribute-a-collection-of-packages-on{to,}-multiple-CDs. Especially good for future use with APT. "distribute" program makes doing the tasks related to creating a CD set for distribution of a collection of packages easier. The tasks include: * laying out the CDs filesystem (splitting the large ... [More] amount of packages into several discs etc.), * preparing the collection for use by APT (indexing), * creating ISO images and recording the discs. Periodical updates to the initially distributed collection can be issued with help of "distribute". This utility will be useful to you, if you have a collection of packages on one machine (say, a mirror), and you want to bring them to your home computer (and install), or if you would like to distribute the collection of packages for some other purpose [Less]

745 lines of code

0 current contributors

over 15 years since last commit

0 users on Open Hub

Inactive
3.0
   
I Use This