# OpenSeq2Seq **Repository Path**: deeplearningrepos/OpenSeq2Seq ## Basic Information - **Project Name**: OpenSeq2Seq - **Description**: Toolkit for efficient experimentation with Speech Recognition, Text2Speech and NLP - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2021-03-30 - **Last Updated**: 2021-08-31 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README [![License](https://img.shields.io/badge/License-Apache%202.0-brightgreen.svg)](https://opensource.org/licenses/Apache-2.0) [![Documentation](https://img.shields.io/badge/documentation-github.io-blue.svg)](https://nvidia.github.io/OpenSeq2Seq/html/index.html)
OpenSeq2Seq
# OpenSeq2Seq: toolkit for distributed and mixed precision training of sequence-to-sequence models OpenSeq2Seq main goal is to allow researchers to most effectively explore various sequence-to-sequence models. The efficiency is achieved by fully supporting distributed and mixed-precision training. OpenSeq2Seq is built using TensorFlow and provides all the necessary building blocks for training encoder-decoder models for neural machine translation, automatic speech recognition, speech synthesis, and language modeling. ## Documentation and installation instructions https://nvidia.github.io/OpenSeq2Seq/ ## Features 1. Models for: 1. Neural Machine Translation 2. Automatic Speech Recognition 3. Speech Synthesis 4. Language Modeling 5. NLP tasks (sentiment analysis) 2. Data-parallel distributed training 1. Multi-GPU 2. Multi-node 3. Mixed precision training for NVIDIA Volta/Turing GPUs ## Software Requirements 1. Python >= 3.5 2. TensorFlow >= 1.10 3. CUDA >= 9.0, cuDNN >= 7.0 4. Horovod >= 0.13 (using Horovod is not required, but is highly recommended for multi-GPU setup) ## Acknowledgments Speech-to-text workflow uses some parts of [Mozilla DeepSpeech](https://github.com/Mozilla/DeepSpeech) project. Beam search decoder with language model re-scoring implementation (in `decoders`) is based on [Baidu DeepSpeech](https://github.com/PaddlePaddle/DeepSpeech). Text-to-text workflow uses some functions from [Tensor2Tensor](https://github.com/tensorflow/tensor2tensor) and [Neural Machine Translation (seq2seq) Tutorial](https://github.com/tensorflow/nmt). ## Disclaimer This is a research project, not an official NVIDIA product. ## Related resources * [Tensor2Tensor](https://github.com/tensorflow/tensor2tensor) * [Neural Machine Translation (seq2seq) Tutorial](https://github.com/tensorflow/nmt) * [OpenNMT](http://opennmt.net/) * [Neural Monkey](https://github.com/ufal/neuralmonkey) * [Sockeye](https://github.com/awslabs/sockeye) * [TF-seq2seq](https://github.com/google/seq2seq) * [Moses](http://www.statmt.org/moses/) ## Paper If you use OpenSeq2Seq, please cite [this paper](https://arxiv.org/abs/1805.10387) ``` @misc{openseq2seq, title={Mixed-Precision Training for NLP and Speech Recognition with OpenSeq2Seq}, author={Oleksii Kuchaiev and Boris Ginsburg and Igor Gitman and Vitaly Lavrukhin and Jason Li and Huyen Nguyen and Carl Case and Paulius Micikevicius}, year={2018}, eprint={1805.10387}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```