-
Notifications
You must be signed in to change notification settings - Fork 8.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ML] File data visualizer: filter running inference endpoints #196577
base: main
Are you sure you want to change the base?
[ML] File data visualizer: filter running inference endpoints #196577
Conversation
Pinging @elastic/ml-ui (:ml) |
…b.com:jgowdyelastic/kibana into file-upload-filter-running-inference-services
…b.com:jgowdyelastic/kibana into file-upload-filter-running-inference-services
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested and LGTM.
One observation - in my local testing, uploading a PDF file (500 lines, 160KB) for the first time can take a couple of minutes - presumably as a new .elser-2
deployment is created. During this time, the progress page stays in step 4 (Uploading data) with no visible sign of anything happening. Is there anything we can do to improve the messaging here to provide more details on what is happening?
I'm now filtering for |
Nice! Thanks |
There will be a change going into es for 8.16 which fixes In the future we could add a step to trigger this auto deployment by calling the infer endpoint during upload and report the status of the download and deployment as an addition step in the upload progress. #196696 |
💛 Build succeeded, but was flaky
Failed CI StepsTest Failures
Metrics [docs]Async chunks
History
|
When listing the inference endpoints available for the semantic text field, we should only list ones where the underlying model is deployed, otherwise the ingest can fail due to a timeout as the model is being deployed and an ML node spun up in the background.
This PR adds a check to the
data_visualizer/inference_endpoints
endpoint to ensure onlysparse_embedding
andtext_embedding
types are used and they have at least one allocation.NOTE, the allocation check is currently commented out waiting on an es change. elastic/elasticsearch#115095
Also renames the endpoint from
data_visualizer/inference_services
->data_visualizer/inference_endpoints
And renames variables which were incorrectly named "service" rather than "endpoint"