This repository has been archived by the owner on Aug 31, 2021. It is now read-only.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
AutoKerasEncoder trains and encodes the documents with custom(best) encoder architecture suiting the dataset, using neural architectural search via AutoKeras.
Data Format: tuple of numpy.ndarray or tf.data.Dataset. The two elements are,
input data - x
for vision (image) : The shape of the data should be should be (samples, width, height) or (samples, width, height, channels).
for bert (text) : The data should be one dimensional. Each element in the data should be a string which is a full sentence.
output data - y (labels)
for classification based training : It can be raw labels, one-hot encoded if more than two classes, or binary encoded for binary classification. The raw labels will be encoded to one column if two classes were found, or one-hot encoded if more than two classes were found.
for regression based training : It can be single-column or multi-column. The values should all be numerical.
model architectures checked and tuned
'vision' mode : ResNet(variants), Xception(variants), conv2d
'bert' mode : Vanilla, Transformer, ngram
Usage:
or,