vmcore: introduce remap_oldmem_pfn_range()

For zfcpdump we can't map the HSA storage because it is only available via
a read interface.  Therefore, for the new vmcore mmap feature we have
introduce a new mechanism to create mappings on demand.

This patch introduces a new architecture function remap_oldmem_pfn_range()
that should be used to create mappings with remap_pfn_range() for oldmem
areas that can be directly mapped.  For zfcpdump this is everything
besides of the HSA memory.  For the areas that are not mapped by
remap_oldmem_pfn_range() a generic vmcore a new generic vmcore fault
handler mmap_vmcore_fault() is called.

This handler works as follows:

* Get already available or new page from page cache (find_or_create_page)
* Check if /proc/vmcore page is filled with data (PageUptodate)
* If yes:
  Return that page
* If no:
  Fill page using __vmcore_read(), set PageUptodate, and return page

Signed-off-by: Michael Holzheu <holzheu@linux.vnet.ibm.com>
Acked-by: Vivek Goyal <vgoyal@redhat.com>
Cc: HATAYAMA Daisuke <d.hatayama@jp.fujitsu.com>
Cc: Jan Willeke <willeke@de.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Michael Holzheu 2013-09-11 14:24:51 -07:00 committed by Linus Torvalds
parent 97b0f6f9cd
commit 9cb218131d
2 changed files with 86 additions and 8 deletions

View File

@ -21,6 +21,7 @@
#include <linux/crash_dump.h> #include <linux/crash_dump.h>
#include <linux/list.h> #include <linux/list.h>
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/pagemap.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
#include <asm/io.h> #include <asm/io.h>
#include "internal.h" #include "internal.h"
@ -153,11 +154,35 @@ ssize_t __weak elfcorehdr_read_notes(char *buf, size_t count, u64 *ppos)
return read_from_oldmem(buf, count, ppos, 0); return read_from_oldmem(buf, count, ppos, 0);
} }
/*
* Architectures may override this function to map oldmem
*/
int __weak remap_oldmem_pfn_range(struct vm_area_struct *vma,
unsigned long from, unsigned long pfn,
unsigned long size, pgprot_t prot)
{
return remap_pfn_range(vma, from, pfn, size, prot);
}
/*
* Copy to either kernel or user space
*/
static int copy_to(void *target, void *src, size_t size, int userbuf)
{
if (userbuf) {
if (copy_to_user((char __user *) target, src, size))
return -EFAULT;
} else {
memcpy(target, src, size);
}
return 0;
}
/* Read from the ELF header and then the crash dump. On error, negative value is /* Read from the ELF header and then the crash dump. On error, negative value is
* returned otherwise number of bytes read are returned. * returned otherwise number of bytes read are returned.
*/ */
static ssize_t read_vmcore(struct file *file, char __user *buffer, static ssize_t __read_vmcore(char *buffer, size_t buflen, loff_t *fpos,
size_t buflen, loff_t *fpos) int userbuf)
{ {
ssize_t acc = 0, tmp; ssize_t acc = 0, tmp;
size_t tsz; size_t tsz;
@ -174,7 +199,7 @@ static ssize_t read_vmcore(struct file *file, char __user *buffer,
/* Read ELF core header */ /* Read ELF core header */
if (*fpos < elfcorebuf_sz) { if (*fpos < elfcorebuf_sz) {
tsz = min(elfcorebuf_sz - (size_t)*fpos, buflen); tsz = min(elfcorebuf_sz - (size_t)*fpos, buflen);
if (copy_to_user(buffer, elfcorebuf + *fpos, tsz)) if (copy_to(buffer, elfcorebuf + *fpos, tsz, userbuf))
return -EFAULT; return -EFAULT;
buflen -= tsz; buflen -= tsz;
*fpos += tsz; *fpos += tsz;
@ -192,7 +217,7 @@ static ssize_t read_vmcore(struct file *file, char __user *buffer,
tsz = min(elfcorebuf_sz + elfnotes_sz - (size_t)*fpos, buflen); tsz = min(elfcorebuf_sz + elfnotes_sz - (size_t)*fpos, buflen);
kaddr = elfnotes_buf + *fpos - elfcorebuf_sz; kaddr = elfnotes_buf + *fpos - elfcorebuf_sz;
if (copy_to_user(buffer, kaddr, tsz)) if (copy_to(buffer, kaddr, tsz, userbuf))
return -EFAULT; return -EFAULT;
buflen -= tsz; buflen -= tsz;
*fpos += tsz; *fpos += tsz;
@ -208,7 +233,7 @@ static ssize_t read_vmcore(struct file *file, char __user *buffer,
if (*fpos < m->offset + m->size) { if (*fpos < m->offset + m->size) {
tsz = min_t(size_t, m->offset + m->size - *fpos, buflen); tsz = min_t(size_t, m->offset + m->size - *fpos, buflen);
start = m->paddr + *fpos - m->offset; start = m->paddr + *fpos - m->offset;
tmp = read_from_oldmem(buffer, tsz, &start, 1); tmp = read_from_oldmem(buffer, tsz, &start, userbuf);
if (tmp < 0) if (tmp < 0)
return tmp; return tmp;
buflen -= tsz; buflen -= tsz;
@ -225,6 +250,55 @@ static ssize_t read_vmcore(struct file *file, char __user *buffer,
return acc; return acc;
} }
static ssize_t read_vmcore(struct file *file, char __user *buffer,
size_t buflen, loff_t *fpos)
{
return __read_vmcore((__force char *) buffer, buflen, fpos, 1);
}
/*
* The vmcore fault handler uses the page cache and fills data using the
* standard __vmcore_read() function.
*
* On s390 the fault handler is used for memory regions that can't be mapped
* directly with remap_pfn_range().
*/
static int mmap_vmcore_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
{
#ifdef CONFIG_S390
struct address_space *mapping = vma->vm_file->f_mapping;
pgoff_t index = vmf->pgoff;
struct page *page;
loff_t offset;
char *buf;
int rc;
page = find_or_create_page(mapping, index, GFP_KERNEL);
if (!page)
return VM_FAULT_OOM;
if (!PageUptodate(page)) {
offset = (loff_t) index << PAGE_CACHE_SHIFT;
buf = __va((page_to_pfn(page) << PAGE_SHIFT));
rc = __read_vmcore(buf, PAGE_SIZE, &offset, 0);
if (rc < 0) {
unlock_page(page);
page_cache_release(page);
return (rc == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS;
}
SetPageUptodate(page);
}
unlock_page(page);
vmf->page = page;
return 0;
#else
return VM_FAULT_SIGBUS;
#endif
}
static const struct vm_operations_struct vmcore_mmap_ops = {
.fault = mmap_vmcore_fault,
};
/** /**
* alloc_elfnotes_buf - allocate buffer for ELF note segment in * alloc_elfnotes_buf - allocate buffer for ELF note segment in
* vmalloc memory * vmalloc memory
@ -271,6 +345,7 @@ static int mmap_vmcore(struct file *file, struct vm_area_struct *vma)
vma->vm_flags &= ~(VM_MAYWRITE | VM_MAYEXEC); vma->vm_flags &= ~(VM_MAYWRITE | VM_MAYEXEC);
vma->vm_flags |= VM_MIXEDMAP; vma->vm_flags |= VM_MIXEDMAP;
vma->vm_ops = &vmcore_mmap_ops;
len = 0; len = 0;
@ -312,9 +387,9 @@ static int mmap_vmcore(struct file *file, struct vm_area_struct *vma)
tsz = min_t(size_t, m->offset + m->size - start, size); tsz = min_t(size_t, m->offset + m->size - start, size);
paddr = m->paddr + start - m->offset; paddr = m->paddr + start - m->offset;
if (remap_pfn_range(vma, vma->vm_start + len, if (remap_oldmem_pfn_range(vma, vma->vm_start + len,
paddr >> PAGE_SHIFT, tsz, paddr >> PAGE_SHIFT, tsz,
vma->vm_page_prot)) vma->vm_page_prot))
goto fail; goto fail;
size -= tsz; size -= tsz;
start += tsz; start += tsz;

View File

@ -17,6 +17,9 @@ extern int __weak elfcorehdr_alloc(unsigned long long *addr,
extern void __weak elfcorehdr_free(unsigned long long addr); extern void __weak elfcorehdr_free(unsigned long long addr);
extern ssize_t __weak elfcorehdr_read(char *buf, size_t count, u64 *ppos); extern ssize_t __weak elfcorehdr_read(char *buf, size_t count, u64 *ppos);
extern ssize_t __weak elfcorehdr_read_notes(char *buf, size_t count, u64 *ppos); extern ssize_t __weak elfcorehdr_read_notes(char *buf, size_t count, u64 *ppos);
extern int __weak remap_oldmem_pfn_range(struct vm_area_struct *vma,
unsigned long from, unsigned long pfn,
unsigned long size, pgprot_t prot);
extern ssize_t copy_oldmem_page(unsigned long, char *, size_t, extern ssize_t copy_oldmem_page(unsigned long, char *, size_t,
unsigned long, int); unsigned long, int);