Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Martin Schwidefsky | 412 | 53.51% | 3 | 20.00% |
Heiko Carstens | 163 | 21.17% | 5 | 33.33% |
Michael Holzheu | 99 | 12.86% | 1 | 6.67% |
Hans-Joachim Picht | 64 | 8.31% | 1 | 6.67% |
Sebastian Ott | 23 | 2.99% | 2 | 13.33% |
Geert Uytterhoeven | 4 | 0.52% | 1 | 6.67% |
Björn Helgaas | 4 | 0.52% | 1 | 6.67% |
Greg Kroah-Hartman | 1 | 0.13% | 1 | 6.67% |
Total | 770 | 15 |
// SPDX-License-Identifier: GPL-2.0 /* * Suspend support specific for s390. * * Copyright IBM Corp. 2009 * * Author(s): Hans-Joachim Picht <hans@linux.vnet.ibm.com> */ #include <linux/pfn.h> #include <linux/suspend.h> #include <linux/mm.h> #include <linux/pci.h> #include <asm/ctl_reg.h> #include <asm/ipl.h> #include <asm/cio.h> #include <asm/sections.h> #include "entry.h" /* * The restore of the saved pages in an hibernation image will set * the change and referenced bits in the storage key for each page. * Overindication of the referenced bits after an hibernation cycle * does not cause any harm but the overindication of the change bits * would cause trouble. * Use the ARCH_SAVE_PAGE_KEYS hooks to save the storage key of each * page to the most significant byte of the associated page frame * number in the hibernation image. */ /* * Key storage is allocated as a linked list of pages. * The size of the keys array is (PAGE_SIZE - sizeof(long)) */ struct page_key_data { struct page_key_data *next; unsigned char data[]; }; #define PAGE_KEY_DATA_SIZE (PAGE_SIZE - sizeof(struct page_key_data *)) static struct page_key_data *page_key_data; static struct page_key_data *page_key_rp, *page_key_wp; static unsigned long page_key_rx, page_key_wx; unsigned long suspend_zero_pages; /* * For each page in the hibernation image one additional byte is * stored in the most significant byte of the page frame number. * On suspend no additional memory is required but on resume the * keys need to be memorized until the page data has been restored. * Only then can the storage keys be set to their old state. */ unsigned long page_key_additional_pages(unsigned long pages) { return DIV_ROUND_UP(pages, PAGE_KEY_DATA_SIZE); } /* * Free page_key_data list of arrays. */ void page_key_free(void) { struct page_key_data *pkd; while (page_key_data) { pkd = page_key_data; page_key_data = pkd->next; free_page((unsigned long) pkd); } } /* * Allocate page_key_data list of arrays with enough room to store * one byte for each page in the hibernation image. */ int page_key_alloc(unsigned long pages) { struct page_key_data *pk; unsigned long size; size = DIV_ROUND_UP(pages, PAGE_KEY_DATA_SIZE); while (size--) { pk = (struct page_key_data *) get_zeroed_page(GFP_KERNEL); if (!pk) { page_key_free(); return -ENOMEM; } pk->next = page_key_data; page_key_data = pk; } page_key_rp = page_key_wp = page_key_data; page_key_rx = page_key_wx = 0; return 0; } /* * Save the storage key into the upper 8 bits of the page frame number. */ void page_key_read(unsigned long *pfn) { struct page *page; unsigned long addr; unsigned char key; page = pfn_to_page(*pfn); addr = (unsigned long) page_address(page); key = (unsigned char) page_get_storage_key(addr) & 0x7f; if (arch_test_page_nodat(page)) key |= 0x80; *(unsigned char *) pfn = key; } /* * Extract the storage key from the upper 8 bits of the page frame number * and store it in the page_key_data list of arrays. */ void page_key_memorize(unsigned long *pfn) { page_key_wp->data[page_key_wx] = *(unsigned char *) pfn; *(unsigned char *) pfn = 0; if (++page_key_wx < PAGE_KEY_DATA_SIZE) return; page_key_wp = page_key_wp->next; page_key_wx = 0; } /* * Get the next key from the page_key_data list of arrays and set the * storage key of the page referred by @address. If @address refers to * a "safe" page the swsusp_arch_resume code will transfer the storage * key from the buffer page to the original page. */ void page_key_write(void *address) { struct page *page; unsigned char key; key = page_key_rp->data[page_key_rx]; page_set_storage_key((unsigned long) address, key & 0x7f, 0); page = virt_to_page(address); if (key & 0x80) arch_set_page_nodat(page, 0); else arch_set_page_dat(page, 0); if (++page_key_rx >= PAGE_KEY_DATA_SIZE) return; page_key_rp = page_key_rp->next; page_key_rx = 0; } int pfn_is_nosave(unsigned long pfn) { unsigned long nosave_begin_pfn = PFN_DOWN(__pa(&__nosave_begin)); unsigned long nosave_end_pfn = PFN_DOWN(__pa(&__nosave_end)); unsigned long end_rodata_pfn = PFN_DOWN(__pa(__end_rodata)) - 1; unsigned long stext_pfn = PFN_DOWN(__pa(_stext)); /* Always save lowcore pages (LC protection might be enabled). */ if (pfn <= LC_PAGES) return 0; if (pfn >= nosave_begin_pfn && pfn < nosave_end_pfn) return 1; /* Skip memory holes and read-only pages (DCSS, ...). */ if (pfn >= stext_pfn && pfn <= end_rodata_pfn) return 0; if (tprot(PFN_PHYS(pfn))) return 1; return 0; } /* * PM notifier callback for suspend */ static int suspend_pm_cb(struct notifier_block *nb, unsigned long action, void *ptr) { switch (action) { case PM_SUSPEND_PREPARE: case PM_HIBERNATION_PREPARE: suspend_zero_pages = __get_free_pages(GFP_KERNEL, LC_ORDER); if (!suspend_zero_pages) return NOTIFY_BAD; break; case PM_POST_SUSPEND: case PM_POST_HIBERNATION: free_pages(suspend_zero_pages, LC_ORDER); break; default: return NOTIFY_DONE; } return NOTIFY_OK; } static int __init suspend_pm_init(void) { pm_notifier(suspend_pm_cb, 0); return 0; } arch_initcall(suspend_pm_init); void save_processor_state(void) { /* swsusp_arch_suspend() actually saves all cpu register contents. * Machine checks must be disabled since swsusp_arch_suspend() stores * register contents to their lowcore save areas. That's the same * place where register contents on machine checks would be saved. * To avoid register corruption disable machine checks. * We must also disable machine checks in the new psw mask for * program checks, since swsusp_arch_suspend() may generate program * checks. Disabling machine checks for all other new psw masks is * just paranoia. */ local_mcck_disable(); /* Disable lowcore protection */ __ctl_clear_bit(0,28); S390_lowcore.external_new_psw.mask &= ~PSW_MASK_MCHECK; S390_lowcore.svc_new_psw.mask &= ~PSW_MASK_MCHECK; S390_lowcore.io_new_psw.mask &= ~PSW_MASK_MCHECK; S390_lowcore.program_new_psw.mask &= ~PSW_MASK_MCHECK; } void restore_processor_state(void) { S390_lowcore.external_new_psw.mask |= PSW_MASK_MCHECK; S390_lowcore.svc_new_psw.mask |= PSW_MASK_MCHECK; S390_lowcore.io_new_psw.mask |= PSW_MASK_MCHECK; S390_lowcore.program_new_psw.mask |= PSW_MASK_MCHECK; /* Enable lowcore protection */ __ctl_set_bit(0,28); local_mcck_enable(); } /* Called at the end of swsusp_arch_resume */ void s390_early_resume(void) { lgr_info_log(); channel_subsystem_reinit(); zpci_rescan(); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1