Expand description
The user-facing JSON web server that listens for inference requests. This is the “front end”. The inference route is automatically created, and distributes inference computation across the array of workers.
Structs
- HTTP request to get all Working workers
- HTTP request to get server statistics
- HTTP request to get the status of all workers