Entropic

Entropic trains and predicts entropy of strings based on character n-gram models. For example:

require 'entropic'
>> m = Entropic::Model.read(open('https://raw.githubusercontent.com/willf/entropy/master/data/google_books_2.tsv')); true
=> true
>> m.predict("entropy")
=> {:log_prob_total=>-37.181802347513745, :log_prob_average=>-6.1969670579189575, :size=>6}
>> m.entropy("yportne")
=> {:log_prob_total=>-34.25705444264748, :log_prob_average=>-5.70950907377458, :size=>6}

The string 'yportne' is much less likely than the string 'entropy'.

You can also train a model, using strings one per line.

>> n = Entropic::Model.new(2); true
=> true
>> File.open('/tmp/training.txt') {|f| n.train(f)}; true
=> true
>> n.entropy('love')
=> 5.132072254636385

You can also train a model, using strings and a count of the number of times it appers, tab separated.

>> o = Entropic::Model.new(2); true
=> true
>> File.open('/tmp/training_with_counts.txt') {|f| o.train_with_multiplier(f)}; true
=> true
>> o.entropy('love')
=> 5.132072254636385

You can also dump a model, to be read later.

>> File.open('/tmp/save.tsv','w') {|f| o.dump(f)}; true
=> true

Installation

Add this line to your application's Gemfile:

gem 'entropic'

And then execute:

$ bundle

Or install it yourself as:

$ gem install entropic

Development

After checking out the repo, run bin/setup to install dependencies. Then, run rake spec to run the tests. You can also run bin/console for an interactive prompt that will allow you to experiment.

To install this gem onto your local machine, run bundle exec rake install. To release a new version, update the version number in version.rb, and then run bundle exec rake release, which will create a git tag for the version, push git commits and tags, and push the .gem file to rubygems.org.

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/willf/entropic.