Tools for data science with a focus on text processing.
- Focuses on "medium data", i.e. data too big to fit into memory but too small to necessitate the use of a cluster.
- Integrates with existing scientific Python stack as well as select outside tools.
- See the
examples/
directory. - The docs contain plots of example output.
- Unix-like command line utilities. Filters (read from stdin/write to stdout) for files
- Wrappers for Python multiprocessing that add ease of use
- Memory-friendly multiprocessing
- Stream text from disk to formats used in common ML processes
- Write processed text to sparse formats
- Helpers for ML tools (e.g. Vowpal Wabbit, Gensim, etc...)
- Other general utilities
- High-level wrappers that have helped with our workflow and provide additional examples of code use
- General ML modeling utilities
Check out the master branch from the rosettarepo. Then, (so long as you have pip
).
cd rosetta
make
make test
Getting the source (above) is the preferred method since the code changes often, but if you don't use Git you can download a tagged release (tarball) here. Then
pip install rosetta-X.X.X.tar.gz
You can check the latest sources with
git clone git://github.com/columbia-applied-data-science/rosetta
Feel free to contribute a bug report or a request by opening an issue
The preferred method to contribute is to fork and send a pull request. Before doing this, read CONTRIBUTING.md
- Major dependencies on Pandas and numpy.
- Minor dependencies on Gensim.
- Some examples need scikit-learn.
From the base repo directory, rosetta/
, you can run all tests with
make test
Documentation is hosted at here. This does NOT auto-update. To make new docs:
cd docs/
make html
Rosetta refers to the Rosetta Stone, the ancient Egyptian tablet discovered just over 200 years ago. The tablet contained fragmented text in three different languages and the uncovering of its meaning is considered an essential key to our understanding of Ancient Egyptian civilization. We would like this project to provide individuals the necessary tools to process and unearth insight in the ever-growing volumes of textual data of today.