You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/content/docs/features/distributed_inferencing.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,7 +11,7 @@ This functionality enables LocalAI to distribute inference requests across multi
11
11
LocalAI supports two modes of distributed inferencing via p2p:
12
12
13
13
-**Federated Mode**: Requests are shared between the cluster and routed to a single worker node in the network based on the load balancer's decision.
14
-
-**Worker Mode**: Requests are processed by all the workers which contributes to the final inference result (by sharing the model weights).
14
+
-**Worker Mode** (aka "model sharding" or "splitting weights"): Requests are processed by all the workers which contributes to the final inference result (by sharing the model weights).
0 commit comments