the-other-tim-brown commented on code in PR #13314:
URL: https://github.com/apache/hudi/pull/13314#discussion_r2094564885
##########
hudi-client/hudi-spark-client/src/main/scala/org/apache/hudi/BaseSparkInternalRowReaderContext.java:
##########
@@ -92,15 +92,16 @@ public Object getValue(InternalRow row, Schema schema,
String fieldName) {
@Override
public HoodieRecord<InternalRow>
constructHoodieRecord(BufferedRecord<InternalRow> bufferedRecord) {
+ HoodieKey hoodieKey = new HoodieKey(bufferedRecord.getRecordKey(),
partitionPath);
if (bufferedRecord.isDelete()) {
return new HoodieEmptyRecord<>(
- new HoodieKey(bufferedRecord.getRecordKey(), null),
+ hoodieKey,
HoodieRecord.HoodieRecordType.SPARK);
}
Schema schema = getSchemaFromBufferRecord(bufferedRecord);
InternalRow row = bufferedRecord.getRecord();
- return new HoodieSparkRecord(row,
HoodieInternalRowUtils.getCachedSchema(schema));
+ return new HoodieSparkRecord(hoodieKey, row,
HoodieInternalRowUtils.getCachedSchema(schema), true);
Review Comment:
@nsivabalan is there a way to know if we will need to copy based on the
instance type of the spark row data?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]