# TextRegressor

### TextRegressor

```
autokeras.TextRegressor(
output_dim=None,
loss="mean_squared_error",
metrics=None,
project_name="text_regressor",
max_trials=100,
directory=None,
objective="val_loss",
tuner=None,
overwrite=False,
seed=None,
max_model_size=None,
**kwargs
)
```

AutoKeras text regression class.

**Arguments**

**output_dim**`int | None`

: Int. The number of output dimensions. Defaults to None. If None, it will be inferred from the data.**loss**`str | Callable | tensorflow.keras.losses.Loss`

: A Keras loss function. Defaults to use 'mean_squared_error'.**metrics**`List[str | Callable | tensorflow.keras.metrics.Metric] | List[List[str | Callable | tensorflow.keras.metrics.Metric]] | Dict[str, str | Callable | tensorflow.keras.metrics.Metric] | None`

: A list of Keras metrics. Defaults to use 'mean_squared_error'.**project_name**`str`

: String. The name of the AutoModel. Defaults to 'text_regressor'.**max_trials**`int`

: Int. The maximum number of different Keras Models to try. The search may finish before reaching the max_trials. Defaults to 100.**directory**`str | pathlib.Path | None`

: String. The path to a directory for storing the search outputs. Defaults to None, which would create a folder with the name of the AutoModel in the current directory.**objective**`str`

: String. Name of model metric to minimize or maximize, e.g. 'val_accuracy'. Defaults to 'val_loss'.**tuner**`str | Type[autokeras.engine.tuner.AutoTuner] | None`

: String or subclass of AutoTuner. If string, it should be one of 'greedy', 'bayesian', 'hyperband' or 'random'. It can also be a subclass of AutoTuner. If left unspecified, it uses a task specific tuner, which first evaluates the most commonly used models for the task before exploring other models.**overwrite**`bool`

: Boolean. Defaults to`False`

. If`False`

, reloads an existing project of the same name if one is found. Otherwise, overwrites the project.**seed**`int | None`

: Int. Random seed.**max_model_size**`int | None`

: Int. Maximum number of scalars in the parameters of a model. Models larger than this are rejected.****kwargs**: Any arguments supported by AutoModel.

### fit

```
TextRegressor.fit(
x=None, y=None, epochs=None, callbacks=None, validation_split=0.2, validation_data=None, **kwargs
)
```

Search for the best model and hyperparameters for the AutoModel.

It will search for the best model based on the performances on validation data.

**Arguments**

**x**: numpy.ndarray or tensorflow.Dataset. Training data x. The input data should be numpy.ndarray or tf.data.Dataset. The data should be one dimensional. Each element in the data should be a string which is a full sentence.**y**: numpy.ndarray or tensorflow.Dataset. Training data y. The targets passing to the head would have to be tf.data.Dataset, np.ndarray, pd.DataFrame or pd.Series. It can be single-column or multi-column. The values should all be numerical.**epochs**: Int. The number of epochs to train each model during the search. If unspecified, by default we train for a maximum of 1000 epochs, but we stop training if the validation loss stops improving for 10 epochs (unless you specified an EarlyStopping callback as part of the callbacks argument, in which case the EarlyStopping callback you specified will determine early stopping).**callbacks**: List of Keras callbacks to apply during training and validation.**validation_split**: Float between 0 and 1. Defaults to 0.2. Fraction of the training data to be used as validation data. The model will set apart this fraction of the training data, will not train on it, and will evaluate the loss and any model metrics on this data at the end of each epoch. The validation data is selected from the last samples in the`x`

and`y`

data provided, before shuffling. This argument is not supported when`x`

is a dataset. The best model found would be fit on the entire dataset including the validation data.**validation_data**: Data on which to evaluate the loss and any model metrics at the end of each epoch. The model will not be trained on this data.`validation_data`

will override`validation_split`

. The type of the validation data should be the same as the training data. The best model found would be fit on the training dataset without the validation data.****kwargs**: Any arguments supported by keras.Model.fit.

**Returns**

history: A Keras History object corresponding to the best model. Its History.history attribute is a record of training loss values and metrics values at successive epochs, as well as validation loss values and validation metrics values (if applicable).

### predict

```
TextRegressor.predict(x, batch_size=32, verbose=1, **kwargs)
```

Predict the output for a given testing data.

**Arguments**

**x**: Any allowed types according to the input node. Testing data.**batch_size**: Number of samples per batch. If unspecified, batch_size will default to 32.**verbose**: Verbosity mode. 0 = silent, 1 = progress bar. Controls the verbosity of keras.Model.predict****kwargs**: Any arguments supported by keras.Model.predict.

**Returns**

A list of numpy.ndarray objects or a single numpy.ndarray. The predicted results.

### evaluate

```
TextRegressor.evaluate(x, y=None, batch_size=32, verbose=1, **kwargs)
```

Evaluate the best model for the given data.

**Arguments**

**x**: Any allowed types according to the input node. Testing data.**y**: Any allowed types according to the head. Testing targets. Defaults to None.**batch_size**: Number of samples per batch. If unspecified, batch_size will default to 32.**verbose**: Verbosity mode. 0 = silent, 1 = progress bar. Controls the verbosity of keras.Model.evaluate****kwargs**: Any arguments supported by keras.Model.evaluate.

**Returns**

Scalar test loss (if the model has a single output and no metrics) or list of scalars (if the model has multiple outputs and/or metrics). The attribute model.metrics_names will give you the display labels for the scalar outputs.

### export_model

```
TextRegressor.export_model()
```

Export the best Keras Model.

**Returns**

keras.Model instance. The best model found during the search, loaded with trained weights.