On 10/25/24 at 05:11pm, David Hildenbrand wrote:
......snip...
> diff --git a/fs/proc/vmcore.c b/fs/proc/vmcore.c
> index 3e90416ee54e..c332a9a4920b 100644
> --- a/fs/proc/vmcore.c
> +++ b/fs/proc/vmcore.c
> @@ -69,6 +69,8 @@ static LIST_HEAD(vmcore_cb_list);
>  /* Whether the vmcore has been opened once. */
>  static bool vmcore_opened;
>  
> +static void vmcore_process_device_ram(struct vmcore_cb *cb);
> +
>  void register_vmcore_cb(struct vmcore_cb *cb)
>  {
>       INIT_LIST_HEAD(&cb->next);
> @@ -80,6 +82,8 @@ void register_vmcore_cb(struct vmcore_cb *cb)
>        */
>       if (vmcore_opened)
>               pr_warn_once("Unexpected vmcore callback registration\n");
> +     else if (cb->get_device_ram)
> +             vmcore_process_device_ram(cb);

Global variable 'vmcore_opened' is used to indicate if /proc/vmcore is
opened. With &vmcore_mutex, we don't need to worry about concurrent
opening and modification. However, if people just open /proc/vmcore and
close it after checking, then s390 will miss the vmcore dumping, is it
acceptable?

>       mutex_unlock(&vmcore_mutex);
>  }
>  EXPORT_SYMBOL_GPL(register_vmcore_cb);
> @@ -1511,6 +1515,158 @@ int vmcore_add_device_dump(struct vmcoredd_data *data)
......
> +
> +static void vmcore_process_device_ram(struct vmcore_cb *cb)
> +{
> +     unsigned char *e_ident = (unsigned char *)elfcorebuf;
> +     struct vmcore_mem_node *first, *m;
> +     LIST_HEAD(list);
> +     int count;
> +
> +     if (cb->get_device_ram(cb, &list)) {
> +             pr_err("Kdump: obtaining device ram ranges failed\n");
> +             return;
> +     }
> +     count = list_count_nodes(&list);
> +     if (!count)
> +             return;
> +
> +     /* We only support Elf64 dumps for now. */
> +     if (WARN_ON_ONCE(e_ident[EI_CLASS] != ELFCLASS64)) {
> +             pr_err("Kdump: device ram ranges only support Elf64\n");
> +             goto out_free;
> +     }

Only supporting Elf64 dumps seems to be a basic checking, do we need
to put it at the beginning of function? Otherwise, we spend efforts to
call cb->get_device_ram(), then fail.

> +
> +     /*
> +      * For some reason these ranges are already know? Might happen
> +      * with unusual register->unregister->register sequences; we'll simply
> +      * sanity check using the first range.
> +      */
> +     first = list_first_entry(&list, struct vmcore_mem_node, list);
> +     list_for_each_entry(m, &vmcore_list, list) {
> +             unsigned long long m_end = m->paddr + m->size;
> +             unsigned long long first_end = first->paddr + first->size;
> +
> +             if (first->paddr < m_end && m->paddr < first_end)
> +                     goto out_free;
> +     }
> +
> +     /* If adding the mem nodes succeeds, they must not be freed. */
> +     if (!vmcore_add_device_ram_elf64(&list, count))
> +             return;
> +out_free:
> +     vmcore_free_mem_nodes(&list);
> +}
> +#else /* !CONFIG_PROC_VMCORE_DEVICE_RAM */
> +static void vmcore_process_device_ram(struct vmcore_cb *cb)
> +{
> +}
> +#endif /* CONFIG_PROC_VMCORE_DEVICE_RAM */
> +
>  /* Free all dumps in vmcore device dump list */
>  static void vmcore_free_device_dumps(void)
>  {
> diff --git a/include/linux/crash_dump.h b/include/linux/crash_dump.h
> index 722dbcff7371..8e581a053d7f 100644
> --- a/include/linux/crash_dump.h
> +++ b/include/linux/crash_dump.h


Reply via email to