diff --git a/output/schema/schema.json b/output/schema/schema.json index 4e7a6eb7ab..8e67cebd2d 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -124976,7 +124976,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L80-L95" + "specLocation": "ml/_types/inference.ts#L85-L100" }, { "kind": "enum", @@ -128271,7 +128271,7 @@ "name": "DeploymentAllocationState", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L280-L293" + "specLocation": "ml/_types/TrainedModel.ts#L282-L295" }, { "kind": "enum", @@ -128293,7 +128293,7 @@ "name": "DeploymentAssignmentState", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L295-L300" + "specLocation": "ml/_types/TrainedModel.ts#L297-L302" }, { "kind": "enum", @@ -128315,7 +128315,7 @@ "name": "DeploymentState", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L265-L278" + "specLocation": "ml/_types/TrainedModel.ts#L267-L280" }, { "kind": "interface", @@ -128814,7 +128814,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L241-L249" + "specLocation": "ml/_types/inference.ts#L256-L264" }, { "kind": "interface", @@ -128860,7 +128860,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L370-L377" + "specLocation": "ml/_types/inference.ts#L395-L402" }, { "kind": "interface", @@ -129066,7 +129066,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L208-L222" + "specLocation": "ml/_types/TrainedModel.ts#L210-L224" }, { "kind": "interface", @@ -129372,6 +129372,19 @@ } } }, + { + "description": "Text expansion configuration for inference.", + "name": "text_expansion", + "required": false, + "since": "8.8.0", + "type": { + "kind": "instance_of", + "type": { + "name": "TextExpansionInferenceOptions", + "namespace": "ml._types" + } + } + }, { "description": "Question answering configuration for inference.", "name": "question_answering", @@ -129386,7 +129399,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L23-L67", + "specLocation": "ml/_types/inference.ts#L23-L72", "variants": { "kind": "container" } @@ -129494,6 +129507,18 @@ } } }, + { + "description": "Text expansion configuration for inference.", + "name": "text_expansion", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "TextExpansionInferenceUpdateOptions", + "namespace": "ml._types" + } + } + }, { "description": "Question answering configuration for inference", "name": "question_answering", @@ -129507,7 +129532,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L265-L285", + "specLocation": "ml/_types/inference.ts#L280-L302", "variants": { "kind": "container" } @@ -129640,7 +129665,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L418-L465" + "specLocation": "ml/_types/inference.ts#L443-L490" }, { "kind": "interface", @@ -131356,7 +131381,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L230-L239" + "specLocation": "ml/_types/inference.ts#L245-L254" }, { "kind": "interface", @@ -131390,7 +131415,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L363-L368" + "specLocation": "ml/_types/inference.ts#L388-L393" }, { "description": "BERT and MPNet tokenization configuration options", @@ -131466,7 +131491,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L116-L143" + "specLocation": "ml/_types/inference.ts#L121-L148" }, { "description": "RoBERTa tokenization configuration options", @@ -131542,7 +131567,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L145-L172" + "specLocation": "ml/_types/inference.ts#L150-L177" }, { "kind": "interface", @@ -131576,7 +131601,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L321-L326" + "specLocation": "ml/_types/inference.ts#L340-L345" }, { "kind": "interface", @@ -131881,7 +131906,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L209-L216" + "specLocation": "ml/_types/inference.ts#L214-L221" }, { "kind": "interface", @@ -131915,7 +131940,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L350-L355" + "specLocation": "ml/_types/inference.ts#L369-L374" }, { "kind": "interface", @@ -131957,7 +131982,7 @@ "name": "PredictedValue", "namespace": "ml._types" }, - "specLocation": "ml/_types/inference.ts#L416-L416", + "specLocation": "ml/_types/inference.ts#L441-L441", "type": { "items": [ { @@ -132049,7 +132074,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L251-L261" + "specLocation": "ml/_types/inference.ts#L266-L276" }, { "kind": "interface", @@ -132119,7 +132144,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L379-L390" + "specLocation": "ml/_types/inference.ts#L404-L415" }, { "kind": "interface", @@ -132156,7 +132181,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L69-L78" + "specLocation": "ml/_types/inference.ts#L74-L83" }, { "kind": "enum", @@ -132186,7 +132211,7 @@ "name": "RoutingState", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L341-L362" + "specLocation": "ml/_types/TrainedModel.ts#L343-L364" }, { "kind": "enum", @@ -132406,7 +132431,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L174-L184" + "specLocation": "ml/_types/inference.ts#L179-L189" }, { "kind": "interface", @@ -132467,7 +132492,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L328-L337" + "specLocation": "ml/_types/inference.ts#L347-L356" }, { "description": "Text embedding inference options", @@ -132477,6 +132502,18 @@ "namespace": "ml._types" }, "properties": [ + { + "description": "The number of dimensions in the embedding output", + "name": "embedding_size", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "integer", + "namespace": "_types" + } + } + }, { "description": "The tokenization options", "name": "tokenization", @@ -132502,7 +132539,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L222-L228" + "specLocation": "ml/_types/inference.ts#L227-L235" }, { "kind": "interface", @@ -132535,7 +132572,75 @@ } } ], - "specLocation": "ml/_types/inference.ts#L357-L361" + "specLocation": "ml/_types/inference.ts#L376-L380" + }, + { + "description": "Text expansion inference options", + "kind": "interface", + "name": { + "name": "TextExpansionInferenceOptions", + "namespace": "ml._types" + }, + "properties": [ + { + "description": "The tokenization options", + "name": "tokenization", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "TokenizationConfigContainer", + "namespace": "ml._types" + } + } + }, + { + "description": "The field that is added to incoming documents to contain the inference prediction. Defaults to predicted_value.", + "name": "results_field", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "string", + "namespace": "_builtins" + } + } + } + ], + "specLocation": "ml/_types/inference.ts#L237-L243" + }, + { + "kind": "interface", + "name": { + "name": "TextExpansionInferenceUpdateOptions", + "namespace": "ml._types" + }, + "properties": [ + { + "name": "tokenization", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "NlpTokenizationUpdateOptions", + "namespace": "ml._types" + } + } + }, + { + "description": "The field that is added to incoming documents to contain the inference prediction. Defaults to predicted_value.", + "name": "results_field", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "string", + "namespace": "_builtins" + } + } + } + ], + "specLocation": "ml/_types/inference.ts#L382-L386" }, { "kind": "interface", @@ -132636,7 +132741,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L97-L114", + "specLocation": "ml/_types/inference.ts#L102-L119", "variants": { "kind": "container" } @@ -132658,7 +132763,7 @@ "name": "TokenizationTruncate", "namespace": "ml._types" }, - "specLocation": "ml/_types/inference.ts#L315-L319" + "specLocation": "ml/_types/inference.ts#L334-L338" }, { "kind": "interface", @@ -132701,7 +132806,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L399-L403" + "specLocation": "ml/_types/inference.ts#L424-L428" }, { "kind": "interface", @@ -132753,7 +132858,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L224-L231" + "specLocation": "ml/_types/TrainedModel.ts#L226-L233" }, { "kind": "interface", @@ -132790,7 +132895,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L233-L238" + "specLocation": "ml/_types/TrainedModel.ts#L235-L240" }, { "kind": "interface", @@ -132836,7 +132941,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L240-L247" + "specLocation": "ml/_types/TrainedModel.ts#L242-L249" }, { "kind": "interface", @@ -132915,7 +133020,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L393-L408" + "specLocation": "ml/_types/TrainedModel.ts#L395-L410" }, { "kind": "interface", @@ -132973,7 +133078,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L364-L382" + "specLocation": "ml/_types/TrainedModel.ts#L366-L384" }, { "kind": "interface", @@ -133079,7 +133184,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L307-L339" + "specLocation": "ml/_types/TrainedModel.ts#L309-L341" }, { "kind": "interface", @@ -133246,9 +133351,9 @@ } }, { - "description": "The default configuration for inference. This can be either a regression, classification, or one of the many NLP focused configurations. It must match the underlying definition.trained_model's target_type.", + "description": "The default configuration for inference. This can be either a regression, classification, or one of the many NLP focused configurations. It must match the underlying definition.trained_model's target_type. For pre-packaged models such as ELSER the config is not required.", "name": "inference_config", - "required": true, + "required": false, "type": { "kind": "instance_of", "type": { @@ -133316,7 +133421,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L157-L191" + "specLocation": "ml/_types/TrainedModel.ts#L159-L193" }, { "kind": "interface", @@ -133341,7 +133446,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L193-L196" + "specLocation": "ml/_types/TrainedModel.ts#L195-L198" }, { "kind": "interface", @@ -133418,7 +133523,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L198-L206" + "specLocation": "ml/_types/TrainedModel.ts#L200-L208" }, { "kind": "interface", @@ -133464,7 +133569,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L384-L391" + "specLocation": "ml/_types/TrainedModel.ts#L386-L393" }, { "kind": "interface", @@ -133636,7 +133741,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L128-L155" + "specLocation": "ml/_types/TrainedModel.ts#L130-L157" }, { "kind": "interface", @@ -133668,6 +133773,18 @@ } } }, + { + "description": "The unique identifier for the trained model deployment.", + "name": "deployment_id", + "required": true, + "type": { + "kind": "instance_of", + "type": { + "name": "Id", + "namespace": "_types" + } + } + }, { "description": "The sum of `error_count` for all nodes in the deployment.", "name": "error_count", @@ -133822,7 +133939,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L62-L97" + "specLocation": "ml/_types/TrainedModel.ts#L62-L99" }, { "kind": "interface", @@ -133887,7 +134004,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L392-L398" + "specLocation": "ml/_types/inference.ts#L417-L423" }, { "kind": "interface", @@ -133919,7 +134036,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L405-L408" + "specLocation": "ml/_types/inference.ts#L430-L433" }, { "kind": "interface", @@ -133965,7 +134082,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L410-L414" + "specLocation": "ml/_types/inference.ts#L435-L439" }, { "kind": "interface", @@ -134037,7 +134154,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L99-L119" + "specLocation": "ml/_types/TrainedModel.ts#L101-L121" }, { "kind": "interface", @@ -134058,7 +134175,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L410-L412" + "specLocation": "ml/_types/TrainedModel.ts#L412-L414" }, { "kind": "interface", @@ -134079,7 +134196,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L414-L416" + "specLocation": "ml/_types/TrainedModel.ts#L416-L418" }, { "kind": "interface", @@ -134113,7 +134230,7 @@ } } ], - "specLocation": "ml/_types/TrainedModel.ts#L121-L126" + "specLocation": "ml/_types/TrainedModel.ts#L123-L128" }, { "kind": "interface", @@ -134226,7 +134343,7 @@ "name": "TrainedModelType", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L249-L263" + "specLocation": "ml/_types/TrainedModel.ts#L251-L265" }, { "kind": "enum", @@ -134242,7 +134359,7 @@ "name": "TrainingPriority", "namespace": "ml._types" }, - "specLocation": "ml/_types/TrainedModel.ts#L302-L305" + "specLocation": "ml/_types/TrainedModel.ts#L304-L307" }, { "kind": "interface", @@ -134349,7 +134466,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L218-L220" + "specLocation": "ml/_types/inference.ts#L223-L225" }, { "description": "Zero shot classification configuration options", @@ -134440,7 +134557,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L186-L207" + "specLocation": "ml/_types/inference.ts#L191-L212" }, { "kind": "interface", @@ -134501,7 +134618,7 @@ } } ], - "specLocation": "ml/_types/inference.ts#L339-L348" + "specLocation": "ml/_types/inference.ts#L358-L367" }, { "attachedBehaviors": [ diff --git a/output/typescript/types.ts b/output/typescript/types.ts index dd8fb973b4..0689adac0b 100644 --- a/output/typescript/types.ts +++ b/output/typescript/types.ts @@ -12475,6 +12475,7 @@ export interface MlInferenceConfigCreateContainer { ner?: MlNerInferenceOptions pass_through?: MlPassThroughInferenceOptions text_embedding?: MlTextEmbeddingInferenceOptions + text_expansion?: MlTextExpansionInferenceOptions question_answering?: MlQuestionAnsweringInferenceOptions } @@ -12487,6 +12488,7 @@ export interface MlInferenceConfigUpdateContainer { ner?: MlNerInferenceUpdateOptions pass_through?: MlPassThroughInferenceUpdateOptions text_embedding?: MlTextEmbeddingInferenceUpdateOptions + text_expansion?: MlTextExpansionInferenceUpdateOptions question_answering?: MlQuestionAnsweringInferenceUpdateOptions } @@ -12803,6 +12805,7 @@ export interface MlTextClassificationInferenceUpdateOptions { } export interface MlTextEmbeddingInferenceOptions { + embedding_size?: integer tokenization?: MlTokenizationConfigContainer results_field?: string } @@ -12812,6 +12815,16 @@ export interface MlTextEmbeddingInferenceUpdateOptions { results_field?: string } +export interface MlTextExpansionInferenceOptions { + tokenization?: MlTokenizationConfigContainer + results_field?: string +} + +export interface MlTextExpansionInferenceUpdateOptions { + tokenization?: MlNlpTokenizationUpdateOptions + results_field?: string +} + export interface MlTimingStats { elapsed_time: DurationValue iteration_time?: DurationValue @@ -12887,7 +12900,7 @@ export interface MlTrainedModelConfig { estimated_heap_memory_usage_bytes?: integer estimated_operations?: integer fully_defined?: boolean - inference_config: MlInferenceConfigCreateContainer + inference_config?: MlInferenceConfigCreateContainer input: MlTrainedModelConfigInput license_level?: string metadata?: MlTrainedModelConfigMetadata @@ -12930,6 +12943,7 @@ export interface MlTrainedModelDeploymentNodesStats { export interface MlTrainedModelDeploymentStats { allocation_status: MlTrainedModelDeploymentAllocationStatus cache_size?: ByteSize + deployment_id: Id error_count: integer inference_count: integer model_id: Id diff --git a/specification/ml/_types/TrainedModel.ts b/specification/ml/_types/TrainedModel.ts index a014ce2231..3e03a3a26a 100644 --- a/specification/ml/_types/TrainedModel.ts +++ b/specification/ml/_types/TrainedModel.ts @@ -63,6 +63,8 @@ export class TrainedModelDeploymentStats { /** The detailed allocation status for the deployment. */ allocation_status: TrainedModelDeploymentAllocationStatus cache_size?: ByteSize + /** The unique identifier for the trained model deployment. */ + deployment_id: Id /** The sum of `error_count` for all nodes in the deployment. */ error_count: integer /** The sum of `inference_count` for all nodes in the deployment. */ @@ -178,8 +180,8 @@ export class TrainedModelConfig { estimated_operations?: integer /** True if the full model definition is present. */ fully_defined?: boolean - /** The default configuration for inference. This can be either a regression, classification, or one of the many NLP focused configurations. It must match the underlying definition.trained_model's target_type. */ - inference_config: InferenceConfigCreateContainer + /** The default configuration for inference. This can be either a regression, classification, or one of the many NLP focused configurations. It must match the underlying definition.trained_model's target_type. For pre-packaged models such as ELSER the config is not required. */ + inference_config?: InferenceConfigCreateContainer /** The input field names for the model definition. */ input: TrainedModelConfigInput /** The license level of the trained model. */ diff --git a/specification/ml/_types/inference.ts b/specification/ml/_types/inference.ts index 347ac04cd3..af48c45e76 100644 --- a/specification/ml/_types/inference.ts +++ b/specification/ml/_types/inference.ts @@ -59,6 +59,11 @@ export class InferenceConfigCreateContainer { * @since 8.0.0 * */ text_embedding?: TextEmbeddingInferenceOptions + /** + * Text expansion configuration for inference. + * @since 8.8.0 + * */ + text_expansion?: TextExpansionInferenceOptions /** * Question answering configuration for inference. * @since 8.3.0 @@ -221,6 +226,16 @@ export class Vocabulary { /** Text embedding inference options */ export class TextEmbeddingInferenceOptions { + /** The number of dimensions in the embedding output */ + embedding_size?: integer + /** The tokenization options */ + tokenization?: TokenizationConfigContainer + /** The field that is added to incoming documents to contain the inference prediction. Defaults to predicted_value. */ + results_field?: string +} + +/** Text expansion inference options */ +export class TextExpansionInferenceOptions { /** The tokenization options */ tokenization?: TokenizationConfigContainer /** The field that is added to incoming documents to contain the inference prediction. Defaults to predicted_value. */ @@ -280,6 +295,8 @@ export class InferenceConfigUpdateContainer { pass_through?: PassThroughInferenceUpdateOptions /** Text embedding configuration for inference. */ text_embedding?: TextEmbeddingInferenceUpdateOptions + /** Text expansion configuration for inference. */ + text_expansion?: TextExpansionInferenceUpdateOptions /** Question answering configuration for inference */ question_answering?: QuestionAnsweringInferenceUpdateOptions } @@ -298,6 +315,8 @@ export class NlpInferenceConfigUpdateContainer { pass_through?: PassThroughInferenceUpdateOptions /** Text embedding configuration for inference. */ text_embedding?: TextEmbeddingInferenceUpdateOptions + /** Text expansion configuration for inference. */ + text_expansion?: TextExpansionInferenceUpdateOptions /** Question answering configuration for inference */ question_answering?: QuestionAnsweringInferenceUpdateOptions } @@ -360,6 +379,12 @@ export class TextEmbeddingInferenceUpdateOptions { results_field?: string } +export class TextExpansionInferenceUpdateOptions { + tokenization?: NlpTokenizationUpdateOptions + /** The field that is added to incoming documents to contain the inference prediction. Defaults to predicted_value. */ + results_field?: string +} + export class NerInferenceUpdateOptions { /** The tokenization options to update when inferring */ tokenization?: NlpTokenizationUpdateOptions