Model description
This is a logistic regression model trained with GPT-2 embeddings on imdb dataset. The notebook to generate this model is in this repository and in this kaggle link.
Intended uses & limitations
This model is trained for educational purposes.
Training Procedure
Hyperparameters
The model is trained with below hyperparameters.
Click to expand
| Hyperparameter | Value |
|---|---|
| memory | |
| steps | [('embedding', HFTransformersLanguage(model_name_or_path='facebook/bart-base')), ('model', LogisticRegression())] |
| verbose | False |
| embedding | HFTransformersLanguage(model_name_or_path='facebook/bart-base') |
| model | LogisticRegression() |
| embedding__model_name_or_path | facebook/bart-base |
| model__C | 1.0 |
| model__class_weight | |
| model__dual | False |
| model__fit_intercept | True |
| model__intercept_scaling | 1 |
| model__l1_ratio | |
| model__max_iter | 100 |
| model__multi_class | auto |
| model__n_jobs | |
| model__penalty | l2 |
| model__random_state | |
| model__solver | lbfgs |
| model__tol | 0.0001 |
| model__verbose | 0 |
| model__warm_start | False |
Model Plot
The model plot is below.
Pipeline(steps=[('embedding',HFTransformersLanguage(model_name_or_path='facebook/bart-base')),('model', LogisticRegression())])Please rerun this cell to show the HTML repr or trust the notebook.Pipeline(steps=[('embedding',HFTransformersLanguage(model_name_or_path='facebook/bart-base')),('model', LogisticRegression())])HFTransformersLanguage(model_name_or_path='facebook/bart-base')
LogisticRegression()
Evaluation Results
You can find the details about evaluation process and the evaluation results.
| Metric | Value |
|---|---|
| f1_score | 0.867535 |
How to Get Started with the Model
Use the code below to get started with the model.
Click to expand
[More Information Needed]
# Additional Content
## Confusion matrix

- Downloads last month
- -