Re: [grpc-io] Re: Inquiry regarding chunking in gRPC

2025-01-28 Thread 'Larry Safran' via grpc.io
You could define your message as a byte array instead of the type you really want to send. Then use ProtoUtils.marshaller(Type.getDefaultInstance()).stream(proto) to turn the proto you are sending into a byte stream which is easy to chunk. Then on the other side, use ProtoUtils.marshaller(Type.ge

[grpc-io] Re: Inquiry regarding chunking in gRPC

2025-01-28 Thread 'Kannan Jayaprakasam' via grpc.io
The 2 GB serialization limit comes protobuf itself so there is nothing we can do about it. You need to write logic to break your layers at possibly arbitrary places not just at layer boundaries, so that each chunk is <= 2GB and regather them on the receiver side with your service code logic. On

[grpc-io] Re: Inquiry regarding chunking in gRPC

2025-01-27 Thread Saurav Pawar
Just a follow up. On Friday, January 24, 2025 at 10:57:51 PM UTC+4 Saurav Pawar wrote: > Hello, hope everything is well with you. > > I am currently using grpc for communication of LLMs (large language > models) having 1 to 7 billion parameters. I know that there is a 2 GB > serialization limit