Skip to content

bioinf-jku/binet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

80 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

binet

binet is a Deep Learning library for Python that was developed by Thomas Unterthiner at the Institute of Bioinformatics of the Johannes Kepler University Linz.

binet can seamlessly and transparently switch between running on the CPU and on the GPU, using PyCUDA and scikits-cuda. It supports dense as well as sparse input data.

The library was written with the goal of easily experimenting with new ideas regarding neural nets. While it is written with high performance in mind, ease of extensibility and to internal net state was the main stated design goal. As a result binet is fast, super flexible and yet also a bit hackish :)

Examples

A simple neural network on MNIST with 2 hidden layers:

import os
from binet import *

op.init_gpu(0)   #  OPTIONAL: initializes first GPU in the system

from binet.util import train
dataset = load_dataset("mnist")

n_inputs = dataset[0].shape[1]
layers = (256, 256, dataset[1].shape[1])
net = NeuralNet(n_inputs, layers, max_iter=10, learning_rate=0.1, verbose=True, \
    activation="relu", shuffle_data=False, dropout=0.5, \
    input_dropout=0.2)
net = train(net, dataset, use_gpu=True, skip_output=1)

Installation

binet requires:

  • numpy
  • scipy
  • pandas
  • cffi
  • Cython
  • h5py (optionally, for load_dataset)
  • PyCUDA
  • scikits.cuda
  • GNU Scientific Library

Citation

If you use binet in a publication and found it useful, please cite

T Unterthiner, A Mayr, G Klambauer, M Steijaert, J Wegner, H Ceulemans, S Hochreiter "Deep Learning as an Opportunity in Virtual Screening" Deep Learning and Representation Learning Workshop (NIPS 2014)

License

binet is licensed under the General Public License (GPL) Version 2 or higher. See License.rst for the full, gory details.

About

Deep Learning package for Python

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published