While widely used in industry, recurrent neural networks (RNNs) are known to have deficiencies dealing with long sequences (e.g. slow inference, vanishing gradients etc.). Recent research has attempted accelerate RNN models by developing mechanisms skip irrelevant words input. Due the lack of labelled data, it remains as a challenge decide which skip, especially for low-resource classification ...