Bert as a Library is a framework for prediction, evaluation and finetuning of Bert models. It's also suitable for production, allowing an easy deploy using Flask or similar services.
You can install the library from pip
using following command
$ pip install BertLibrary
The setup of BertLibrary is dead simple. You have 2 options:
import BertFTModel
to finetune your model and run evaluations/predictions.import BertFEModel
if you want to extract features from a pretrained/finetuned Bert (only prediction/evaluation).- Download a pretrained model from here
# Import model for fintuning
from BertLibary import BertFTModel
# Instantiate the model
ft_model = BertFTModel( model_dir='uncased_L-12_H-768_A-12',
ckpt_name="bert_model.ckpt",
labels=['0','1'],
lr=1e-05,
num_train_steps=30000,
num_warmup_steps=1000,
ckpt_output_dir='output',
save_check_steps=1000,
do_lower_case=False,
max_seq_len=50,
batch_size=32,
)
ft_trainer = ft_model.get_trainer()
ft_predictor = ft_model.get_predictor()
ft_evaluator = ft_model.get_evaluator()
BertFTModel constructor parameters:
Command | Description |
---|---|
model_dir | The path to the Bert pretrained model directory |
ckpt_name | The name of the checkpoint you want use |
labels | The list of unique label names (must be string) |
lr | The learning rate you will use during the finetuning |
num_train_steps | The default number of steps to run the finetuning if not specified |
num_warmup_steps | Number of warmup steps, see the original paper for more |
ckpt_output_dir | The directory to save the finetuned model checkpoints |
save_check_steps | Save and evaluate the model every save_check_steps |
do_lower_case | Do a lower case during preprocessing if set to true |
max_seq_len | Set a max sequence length of the model (max 512) |
batch_size | Regulate the batch size for training/evaluation/prediction |
config | (Optional) Tensorflow config object |
# Import model for doing feature extraction
from BertLibrary import BertFEModel
# Instantiate the model
fe_model = BertFEModel( model_dir='uncased_L-12_H-768_A-12',
ckpt_name="bert_model.ckpt",
layer=-2,
do_lower_case=False,
max_seq_len=50,
batch_size=32,
)
fe_predictor = fe_model.get_predictor()
BertFEModel constructor parameters:
Command | Description |
---|---|
model_dir | The path to the Bert pretrained model directory |
ckpt_name | The name of the checkpoint you want use |
layer | The number of layer to select for feature extractions. You can use negative indexes like -2 |
do_lower_case | Do a lower case during preprocessing if set to true |
max_seq_len | Set a max sequence length of the model (max 512) |
batch_size | Regulate the batch size for training/evaluation/prediction |
config | (Optional) Tensorflow config object |
To run the training you have 2 options: run from file and run from memory.
If you want to run from file you must first create 3 separate files (first two for training, last one for evaluation) which are train|dev|test.tsv
. All those files must respect the tsv format having as first column the string label and the second column the text
with no header.
Example:
Remember to | not inset the header |
---|---|
positive | I love cats |
negative | I hate cucumbers |
For example if you have a folder ./dataset
with train.tsv
and dev.tsv
, you can run the finetuning like this:
path_to_file = './datasets'
ft_trainer.train_from_file(path_to_file, 60000)
Instead if you want to run the finetuning from memory, you can pass just the data like in the following example where X
and X_val
are a list of sentences, and y
contains the string labels
X, y, X_val, y_val = get_train_test_data()
ft_trainer.train(X, y, X_val=X_val, y_val=y_val, 60000)
To run evaluation you can run this command if you are working with files
ft_evaluator.evaluate_from_file(path_to_file, checkpoint="output/model.ckpt-35000")
Or you can run it like this on memory
ft_evaluator.evaluate(X_test, y_test, checkpoint="output/model.ckpt-35000")
To run predictions you must instantiate the model first, and then you get the predictor using get_predictor funciton from the model object
# Get the predictor from the desired model
predictor = model.get_predictor()
# Call the predictor passing a list of sentences
predictor(sentences)