Brief Review — Bidirectional Recurrent Neural Networks

Bidirectional RNN (BRNN), One Forward, One Backward

  • A regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN), by training it simultaneously in positive and negative time direction.

Outline

  1. Bidirectional Recurrent Neural Network (BRNN)
  2. Results

1. Bidirectional Recurrent Neural Network (BRNN)

1. Conventional Uni-Directional RNN

General structure of a regular unidirectional RNN shown (a) with a delay line and (b) unfolded in time for two time steps
  • The above figure shows a basic RNN architecture with a delay line and unfolded in time for two time steps.

1.2. Proposed Bidirectional RNN (BRNN)

General structure of the bidirectional recurrent neural network (BRNN) shown unfolded in time for three time steps

2. Results

TIMIT Phoneme Classification Results
  • The TIMIT phoneme database is a well-established database consisting of 6300 sentences spoken by 630 speakers.
  • The training data set consisting of 3696 sentences from 462 speakers, and the test data set consisting of 1344 sentences from 168 speakers.
  • The data is segment which gives 142910 phoneme segments for training and 51681 for testing.
  • MERGE is the method to merge results from FOR-RNN and BACK-RNN.
Modified bidirectional recurrent neural network structure shown here with extensions for the forward posterior probability estimation
  • A more enhanced BRNN is proposed.
  • First, instead of connecting the forward and backward states to the current output states, they are connected to the next and previous output states, respectively, and the inputs are directly connected to the outputs.
  • Second, if in the resulting structure the first L weight connections from the inputs to the backward states and the inputs to the outputs are cut, then only discrete input information from can be used to make predictions.
TIMIT Phoneme Classification Results

Reference

[1997 TSP] [Bidirectional RNN (BRNN)]
Bidirectional Recurrent Neural Networks

Language Model / Sequence Model

1997 [Bidirectional RNN (BRNN)] 2020 [ALBERT] [GPT-3] [T5] [Pre-LN Transformer] [MobileBERT] [TinyBERT]

My Other Previous Paper Readings

--

--

PhD, Researcher. I share what I learn. :) Linktree: https://linktr.ee/shtsang for Twitter, LinkedIn, etc.

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store