Model Inference (experimental)
Example showing how to use the inference functionality
Last updated
Example showing how to use the inference functionality
Last updated
The Bubbaloop server is able to run inference efficiently Visual Language Models (VLM) using directly the camera streams without any latency in the same process and broadcast the results.
Supported models (via / )
PaliGemma:
Similar to the pipeline, we can customize the inference.ron
pipeline to adjust to our system setup. This will require compiling every time you modify your config.
By default, this command will start the inference engine using the prompt "cap en" — to generate a short capture from each frame.
In your terminal you should be able to get somethin similar
We expose some setting via a REST api to the following end point.
This will fix the prompt to run inference on to detect people
You can access also to the image streams and prompts results via the following API including their timestamps.
To stop the pipeline, use the stop-pipeline
command:
Check the supported prompts:
We provide a small Python script that calls the above end points and visualize the results with