-
Notifications
You must be signed in to change notification settings - Fork 167
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Convert hivemind.Server/RemoteModuleCall/RemoteCallMany to libp2p backend #242
Labels
Comments
working on it |
Current status:
On merging:
Current affairs
|
Current progress:
|
Current status: Implemented streaming for backward and forward for moe/expert/RemoteModuleCall. Now we split all inputs and outputs for forward/backward into smaller parts before sending, then stream this parts and assemble again on the other end. There is no problem with big batches now. In plans:
|
Current status:
In plans:
|
15 tasks
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
[depends on #238 to be merged ]
After we've implemented P2P transport with nat traversal, we should switch the main components to libp2p backend to take advantage of this new transport.
One of three main components is
hivemind.server.Server
and its counterparthivemind.client.RemoteExpert
On a client side, hivemind creates a RemoteExpert pytorch module that calls experts via
_RemoteModuleCall
(and_RemoteCallMany
for DMoE)A server receives incoming connections with several
ConnectionHandler
processes running in parallel. These processes run gRPC servers and hence should be switched to libp2p.tests/test_moe.py
tests/test_training.py
tests/benchmark_througphput.py
The text was updated successfully, but these errors were encountered: