Attenders: Yiyang, Xi(Pony), Hao, Minyu, Bin, Hualong, Yuanben, Jianghua,
Conway, Jeff, Wei-Chiu, Sammi

Cloudera:

   - Share Ozone append initial design.
   - Present the web site under development.
   - Suggest community to pay attention to the "Ozone Storage Container
   Reconciliation" design and better to have a review.

DiDi:

   - Suggest one improvement on Recon UI, fold the pipeline info for a DN
   on Datanode page. When a Datanode has more than 10 pipelines, it will
   display all the pipelines by default, which is not necessary most of the
   time. @devmadhuu <https://github.com/devmadhuu>
   - OM transaction ID crash bug fixed in HDDS-9876
   <https://issues.apache.org/jira/browse/HDDS-9876>. One more further step
   to help to recover the OM DB, would be a CLI tool to update the transaction
   id in RocksDB. Filed HDDS-10295
   <https://issues.apache.org/jira/browse/HDDS-10295>
   - Testing Follower reader performance. One observation about Follower
   read, the client always reads from the first SCM defined in the HA group.
   It's better to randomly select some SCM to start with, so the read request
   will be served evenly across SCM instances. @szetszwo
   <https://github.com/szetszwo> .
   - Reported one block missing issue. After a further investigation after
   the meeting, it is caused by some customized lock changes on the DN side.
   After reverting the changes, the issue should be fixed.

Shopee:

   - Discussed unsupport HCFS APIs in Ozone
      - append, truncate
      - access time
      - EC policy change API
      - no directory rename for open for writing files.
   - Discussed the way to share files between different users.
      - S3 has such capability called presigned URLs. Ozone should have
      supported this too.
      - symbolic link is not supported in Ozone so far. And there is no
      plan to support it yet.

Qihoo:

   - Find one performance issue with volume check, if the disk is not
   physically healthy, the volume check process which tries to write and read
   a small amount of data file will itself hang for around 1 hour, then the
   disk cannot be marked as failed timely. During this 1h, many read/write
   operations scheduled on this disk, causing the whole DN read/write latency
   extremely high. The solution is found and will submit a patch later.
   - DELETED state containers in memory. There are many DELETED state
   container info held in memory after a bunch of data writes, then deleted
   shortly after that. This data in memory will never be freed, nor will it be
   deleted from SCM rocksDB table. Propose a lazy way to delete these
   containers from memory and rocksDB. Will have a simple design document and
   share with the community.
   - Block location cache, which is not cleared and so a client always
   fetches from the same DN for the same block.
   (XceiverClientGrpc.getBlockDNcache). This will be a problem for long
   running services like S3 gateway, which usually holds the client for its
   entire life. So some DNs will become read hot spot.

Reply via email to