danny0405 commented on code in PR #13699:
URL: https://github.com/apache/hudi/pull/13699#discussion_r2273598475


##########
hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/HoodieTable.java:
##########
@@ -150,6 +151,7 @@ public abstract class HoodieTable<T, I, K, O> implements 
Serializable {
 
   private transient FileSystemViewManager viewManager;
   protected final transient HoodieEngineContext context;
+  private final ReaderContextFactory<T> readerContextFactoryForWrite;

Review Comment:
   The hoodie table holds all the info required there: the engine context, the 
meta client, and the write config. It looks not right we pass around a "reader" 
context factory for all the write path, even for write path that is not COW 
table merging scenarios.
   
   There are also discrepancies in write path, only spark needs this factory, 
Flink and Java can get the factory directly from the engine context in the 
hoodie table.
   
   The broadcast is already there for all the write executors before my change, 
if we have some manner to resolve the serialization issue of the engine context 
or reader context itself, that would be the best. Or if we can limit the 
factory only been initialized for COW table updates in write handles, that 
would be great.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to