runAnalyzer()
This operation processes the input data and generates tokenized output.
public RunAnalyzerResp runAnalyzer(RunAnalyzerReq request)
Request Syntax
runAnalyzer(RunAnalyzerReq.builder()
.texts(List<String> texts)
.analyzerParams(Map<String, Object> analyzerParams)
.withDetail(Boolean withDetail)
.withHash(Boolean withHash)
.databaseName(String databaseName)
.collectionName(String collectionName)
.fieldName(String fieldName)
.analyzerNames(List<String> analyzerNames)
.build()
);
BUILDER METHODS:
-
texts(List<String> texts)-A list of text strings to analyze.
-
analyzerParams(Map<String, Object> analyzerParams)-A map of analyzer parameters.
-
withDetail(Boolean withDetail)-Whether to include detailed token information.
-
withHash(Boolean withHash)-Whether to include hash values in the output.
-
databaseName(String databaseName)-The name of the database. Defaults to the current database if not specified.
-
collectionName(String collectionName)-The name of the target collection.
-
fieldName(String fieldName)-The name of the target field.
-
analyzerNames(List<String> analyzerNames)-A list of analyzer names to use.
RETURNS:
RunAnalyzerResp
A RunAnalyzerResp contains a list of AnalyzerResult objects, each of which is a list of AnalyzerToken objects.
EXCEPTIONS:
-
MilvusClientException
This exception will be raised when any error occurs during this operation.
Example
import io.milvus.v2.client.ConnectConfig;
import io.milvus.v2.client.MilvusClientV2;
import io.milvus.v2.service.vector.request.RunAnalyzerReq;
import io.milvus.v2.service.vector.response.RunAnalyzerResp;
// 1. Set up a client
ConnectConfig connectConfig = ConnectConfig.builder()
.uri("YOUR_CLUSTER_ENDPOINT")
.token("YOUR_CLUSTER_TOKEN")
.build();
MilvusClientV2 client = new MilvusClientV2(connectConfig);
// 2. Run analyzer
List<String> texts = new ArrayList<>();
texts.add("Analyzers (tokenizers) for multi languages");
texts.add("2.5 to take advantage of enhancements and fixes!");
Map<String, Object> analyzerParams = new HashMap<>();
analyzerParams.put("tokenizer", "standard");
analyzerParams.put("filter",
Arrays.asList("lowercase",
new HashMap<String, Object>() {{
put("type", "stop");
put("stop_words", Arrays.asList("to", "of", "for", "the"));
}}));
RunAnalyzerResp resp = client.runAnalyzer(RunAnalyzerReq.builder()
.texts(texts)
.analyzerParams(analyzerParams)
.withDetail(true)
.withHash(true)
.build());