>> /*
>>  * For now, we have a linear search to go find the appropriate
>>  * memory_block corresponding to a particular phys_index. If
>> @@ -658,6 +670,11 @@ static int init_memory_block(struct memory_block 
>> **memory, int block_id,
>>      unsigned long start_pfn;
>>      int ret = 0;
>>
>> +    mem = find_memory_block_by_id(block_id, NULL);
>> +    if (mem) {
>> +            put_device(&mem->dev);
>> +            return -EEXIST;
>> +    }
> 
> find_memory_block_by_id() is not that close to the main idea in this patch.
> Would it be better to split this part?

I played with that but didn't like the temporary results (e.g. having to
export find_memory_block_by_id()). I'll stick to this for now.

> 
>>      mem = kzalloc(sizeof(*mem), GFP_KERNEL);
>>      if (!mem)
>>              return -ENOMEM;
>> @@ -699,44 +716,53 @@ static int add_memory_block(int base_section_nr)
>>      return 0;
>> }
>>
>> +static void unregister_memory(struct memory_block *memory)
>> +{
>> +    if (WARN_ON_ONCE(memory->dev.bus != &memory_subsys))
>> +            return;
>> +
>> +    /* drop the ref. we got via find_memory_block() */
>> +    put_device(&memory->dev);
>> +    device_unregister(&memory->dev);
>> +}
>> +
>> /*
>> - * need an interface for the VM to add new memory regions,
>> - * but without onlining it.
>> + * Create memory block devices for the given memory area. Start and size
>> + * have to be aligned to memory block granularity. Memory block devices
>> + * will be initialized as offline.
>>  */
>> -int hotplug_memory_register(int nid, struct mem_section *section)
>> +int create_memory_block_devices(unsigned long start, unsigned long size)
>> {
>> -    int block_id = base_memory_block_id(__section_nr(section));
>> -    int ret = 0;
>> +    const int start_block_id = pfn_to_block_id(PFN_DOWN(start));
>> +    int end_block_id = pfn_to_block_id(PFN_DOWN(start + size));
>>      struct memory_block *mem;
>> +    unsigned long block_id;
>> +    int ret = 0;
>>
>> -    mutex_lock(&mem_sysfs_mutex);
>> +    if (WARN_ON_ONCE(!IS_ALIGNED(start, memory_block_size_bytes()) ||
>> +                     !IS_ALIGNED(size, memory_block_size_bytes())))
>> +            return -EINVAL;
>>
>> -    mem = find_memory_block(section);
>> -    if (mem) {
>> -            mem->section_count++;
>> -            put_device(&mem->dev);
>> -    } else {
>> +    mutex_lock(&mem_sysfs_mutex);
>> +    for (block_id = start_block_id; block_id != end_block_id; block_id++) {
>>              ret = init_memory_block(&mem, block_id, MEM_OFFLINE);
>>              if (ret)
>> -                    goto out;
>> -            mem->section_count++;
>> +                    break;
>> +            mem->section_count = sections_per_block;
>> +    }
>> +    if (ret) {
>> +            end_block_id = block_id;
>> +            for (block_id = start_block_id; block_id != end_block_id;
>> +                 block_id++) {
>> +                    mem = find_memory_block_by_id(block_id, NULL);
>> +                    mem->section_count = 0;
>> +                    unregister_memory(mem);
>> +            }
>>      }
> 
> Would it be better to do this in reverse order?
> 
> And unregister_memory() would free mem, so it is still necessary to set
> section_count to 0?

1. I kept the existing behavior (setting it to 0) for now. I am planning
to eventually remove the section count completely (it could be
beneficial to detect removing of partially populated memory blocks).

2. Reverse order: We would have to start with "block_id - 1", I don't
like that better.

Thanks for having a look!

-- 

Thanks,

David / dhildenb

Reply via email to