Coder Social home page Coder Social logo

lvapeab / multimodal_keras_wrapper Goto Github PK

View Code? Open in Web Editor NEW

This project forked from marcbs/multimodal_keras_wrapper

9.0 2.0 7.0 102.02 MB

Wrapper for Keras with support to easy data loading and handling and the creation of staged networks.

License: MIT License

Jupyter Notebook 71.02% Python 28.42% Makefile 0.32% HTML 0.02% C++ 0.01% Cuda 0.20% Shell 0.01%

multimodal_keras_wrapper's Introduction

Multimodal Keras Wrapper

Wrapper for Keras with support to easy multimodal data and models loading and handling.

PyPI version Build Status Requirements Status Compatibility license

Documentation

You can access the library documentation page at marcbs.github.io/multimodal_keras_wrapper/

Some code examples are available in demo.ipynb and test.py. Additionally, in the section Projects you can see some practical examples of projects using this library.

Installation

Assuming that you have pip installed, run:

pip install multimodal-keras-wrapper

Alternatively, if you want to install the library from the source code, you just have to follow these steps:

  1. Clone this repository.

  2. Include the repository path into your PYTHONPATH:

export PYTHONPATH=$PYTHONPATH:/path/to/multimodal_keras_wrapper
  1. Install the dependencies (it will install our custom Keras fork):
pip install -r requirements.txt

Additional dependencies

The following additional dependencies are required to fully exploit this library:

  • Keras - custom fork or original version
  • The cupy package can be used for performing numpy-like operations in the GPU. If not available, the package will fall back to numpy.
  • Coco-caption evaluation package (Only required to perform COCO evaluation). This package requires java (version 1.8.0 or newer).

Only when using NMS for certain localization utilities:

Projects

You can see more practical examples in projects which use this library:

TMA for Egocentric Video Captioning based on Temporally-linked Sequences.

NMT-Keras: Neural Machine Translation.

VIBIKNet for Visual Question Answering

ABiViRNet for Video Description

Sentence-SelectioNN for Domain Adaptation in SMT

Keras

For additional information on the Deep Learning library, visit the official web page www.keras.io or the GitHub repository https://github.com/keras-team/keras.

You can also use our custom Keras version, which provides several additional layers for Multimodal Learning.

multimodal_keras_wrapper's People

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.