Merge branch 'kmap_atomic' of git://github.com/congwang/linux
Pull final kmap_atomic cleanups from Cong Wang: "This should be the final round of cleanup, as the definitions of enum km_type finally get removed from the whole tree. The patches have been in linux-next for a long time." * 'kmap_atomic' of git://github.com/congwang/linux: pipe: remove KM_USER0 from comments vmalloc: remove KM_USER0 from comments feature-removal-schedule.txt: remove kmap_atomic(page, km_type) tile: remove km_type definitions um: remove km_type definitions asm-generic: remove km_type definitions avr32: remove km_type definitions frv: remove km_type definitions powerpc: remove km_type definitions arm: remove km_type definitions highmem: remove the deprecated form of kmap_atomic tile: remove usage of enum km_type frv: remove the second parameter of kmap_atomic_primary() jbd2: remove the second argument of kmap_atomic
This commit is contained in:
commit
84eda28060
18 changed files with 44 additions and 283 deletions
|
@ -512,14 +512,6 @@ Who: Sebastian Andrzej Siewior <sebastian@breakpoint.cc>
|
||||||
|
|
||||||
----------------------------
|
----------------------------
|
||||||
|
|
||||||
What: kmap_atomic(page, km_type)
|
|
||||||
When: 3.5
|
|
||||||
Why: The old kmap_atomic() with two arguments is deprecated, we only
|
|
||||||
keep it for backward compatibility for few cycles and then drop it.
|
|
||||||
Who: Cong Wang <amwang@redhat.com>
|
|
||||||
|
|
||||||
----------------------------
|
|
||||||
|
|
||||||
What: get_robust_list syscall
|
What: get_robust_list syscall
|
||||||
When: 2013
|
When: 2013
|
||||||
Why: There appear to be no production users of the get_robust_list syscall,
|
Why: There appear to be no production users of the get_robust_list syscall,
|
||||||
|
|
|
@ -4,30 +4,6 @@
|
||||||
/*
|
/*
|
||||||
* This is the "bare minimum". AIO seems to require this.
|
* This is the "bare minimum". AIO seems to require this.
|
||||||
*/
|
*/
|
||||||
enum km_type {
|
#define KM_TYPE_NR 16
|
||||||
KM_BOUNCE_READ,
|
|
||||||
KM_SKB_SUNRPC_DATA,
|
|
||||||
KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KM_USER0,
|
|
||||||
KM_USER1,
|
|
||||||
KM_BIO_SRC_IRQ,
|
|
||||||
KM_BIO_DST_IRQ,
|
|
||||||
KM_PTE0,
|
|
||||||
KM_PTE1,
|
|
||||||
KM_IRQ0,
|
|
||||||
KM_IRQ1,
|
|
||||||
KM_SOFTIRQ0,
|
|
||||||
KM_SOFTIRQ1,
|
|
||||||
KM_L1_CACHE,
|
|
||||||
KM_L2_CACHE,
|
|
||||||
KM_KDB,
|
|
||||||
KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
#ifdef CONFIG_DEBUG_HIGHMEM
|
|
||||||
#define KM_NMI (-1)
|
|
||||||
#define KM_NMI_PTE (-1)
|
|
||||||
#define KM_IRQ_PTE (-1)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -2,29 +2,9 @@
|
||||||
#define __ASM_AVR32_KMAP_TYPES_H
|
#define __ASM_AVR32_KMAP_TYPES_H
|
||||||
|
|
||||||
#ifdef CONFIG_DEBUG_HIGHMEM
|
#ifdef CONFIG_DEBUG_HIGHMEM
|
||||||
# define D(n) __KM_FENCE_##n ,
|
# define KM_TYPE_NR 29
|
||||||
#else
|
#else
|
||||||
# define D(n)
|
# define KM_TYPE_NR 14
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
enum km_type {
|
|
||||||
D(0) KM_BOUNCE_READ,
|
|
||||||
D(1) KM_SKB_SUNRPC_DATA,
|
|
||||||
D(2) KM_SKB_DATA_SOFTIRQ,
|
|
||||||
D(3) KM_USER0,
|
|
||||||
D(4) KM_USER1,
|
|
||||||
D(5) KM_BIO_SRC_IRQ,
|
|
||||||
D(6) KM_BIO_DST_IRQ,
|
|
||||||
D(7) KM_PTE0,
|
|
||||||
D(8) KM_PTE1,
|
|
||||||
D(9) KM_PTE2,
|
|
||||||
D(10) KM_IRQ0,
|
|
||||||
D(11) KM_IRQ1,
|
|
||||||
D(12) KM_SOFTIRQ0,
|
|
||||||
D(13) KM_SOFTIRQ1,
|
|
||||||
D(14) KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
#undef D
|
|
||||||
|
|
||||||
#endif /* __ASM_AVR32_KMAP_TYPES_H */
|
#endif /* __ASM_AVR32_KMAP_TYPES_H */
|
||||||
|
|
|
@ -76,15 +76,16 @@ extern struct page *kmap_atomic_to_page(void *ptr);
|
||||||
|
|
||||||
#ifndef __ASSEMBLY__
|
#ifndef __ASSEMBLY__
|
||||||
|
|
||||||
#define __kmap_atomic_primary(type, paddr, ampr) \
|
#define __kmap_atomic_primary(cached, paddr, ampr) \
|
||||||
({ \
|
({ \
|
||||||
unsigned long damlr, dampr; \
|
unsigned long damlr, dampr; \
|
||||||
\
|
\
|
||||||
dampr = paddr | xAMPRx_L | xAMPRx_M | xAMPRx_S | xAMPRx_SS_16Kb | xAMPRx_V; \
|
dampr = paddr | xAMPRx_L | xAMPRx_M | xAMPRx_S | xAMPRx_SS_16Kb | xAMPRx_V; \
|
||||||
\
|
\
|
||||||
if (type != __KM_CACHE) \
|
if (!cached) \
|
||||||
asm volatile("movgs %0,dampr"#ampr :: "r"(dampr) : "memory"); \
|
asm volatile("movgs %0,dampr"#ampr :: "r"(dampr) : "memory"); \
|
||||||
else \
|
else \
|
||||||
|
/* cache flush page attachment point */ \
|
||||||
asm volatile("movgs %0,iampr"#ampr"\n" \
|
asm volatile("movgs %0,iampr"#ampr"\n" \
|
||||||
"movgs %0,dampr"#ampr"\n" \
|
"movgs %0,dampr"#ampr"\n" \
|
||||||
:: "r"(dampr) : "memory" \
|
:: "r"(dampr) : "memory" \
|
||||||
|
@ -112,29 +113,20 @@ extern struct page *kmap_atomic_to_page(void *ptr);
|
||||||
(void *) damlr; \
|
(void *) damlr; \
|
||||||
})
|
})
|
||||||
|
|
||||||
static inline void *kmap_atomic_primary(struct page *page, enum km_type type)
|
static inline void *kmap_atomic_primary(struct page *page)
|
||||||
{
|
{
|
||||||
unsigned long paddr;
|
unsigned long paddr;
|
||||||
|
|
||||||
pagefault_disable();
|
pagefault_disable();
|
||||||
paddr = page_to_phys(page);
|
paddr = page_to_phys(page);
|
||||||
|
|
||||||
switch (type) {
|
return __kmap_atomic_primary(1, paddr, 2);
|
||||||
case 0: return __kmap_atomic_primary(0, paddr, 2);
|
|
||||||
case 1: return __kmap_atomic_primary(1, paddr, 3);
|
|
||||||
case 2: return __kmap_atomic_primary(2, paddr, 4);
|
|
||||||
case 3: return __kmap_atomic_primary(3, paddr, 5);
|
|
||||||
|
|
||||||
default:
|
|
||||||
BUG();
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#define __kunmap_atomic_primary(type, ampr) \
|
#define __kunmap_atomic_primary(cached, ampr) \
|
||||||
do { \
|
do { \
|
||||||
asm volatile("movgs gr0,dampr"#ampr"\n" ::: "memory"); \
|
asm volatile("movgs gr0,dampr"#ampr"\n" ::: "memory"); \
|
||||||
if (type == __KM_CACHE) \
|
if (cached) \
|
||||||
asm volatile("movgs gr0,iampr"#ampr"\n" ::: "memory"); \
|
asm volatile("movgs gr0,iampr"#ampr"\n" ::: "memory"); \
|
||||||
} while(0)
|
} while(0)
|
||||||
|
|
||||||
|
@ -143,17 +135,9 @@ do { \
|
||||||
asm volatile("tlbpr %0,gr0,#4,#1" : : "r"(vaddr) : "memory"); \
|
asm volatile("tlbpr %0,gr0,#4,#1" : : "r"(vaddr) : "memory"); \
|
||||||
} while(0)
|
} while(0)
|
||||||
|
|
||||||
static inline void kunmap_atomic_primary(void *kvaddr, enum km_type type)
|
static inline void kunmap_atomic_primary(void *kvaddr)
|
||||||
{
|
{
|
||||||
switch (type) {
|
__kunmap_atomic_primary(1, 2);
|
||||||
case 0: __kunmap_atomic_primary(0, 2); break;
|
|
||||||
case 1: __kunmap_atomic_primary(1, 3); break;
|
|
||||||
case 2: __kunmap_atomic_primary(2, 4); break;
|
|
||||||
case 3: __kunmap_atomic_primary(3, 5); break;
|
|
||||||
|
|
||||||
default:
|
|
||||||
BUG();
|
|
||||||
}
|
|
||||||
pagefault_enable();
|
pagefault_enable();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2,28 +2,6 @@
|
||||||
#ifndef _ASM_KMAP_TYPES_H
|
#ifndef _ASM_KMAP_TYPES_H
|
||||||
#define _ASM_KMAP_TYPES_H
|
#define _ASM_KMAP_TYPES_H
|
||||||
|
|
||||||
enum km_type {
|
#define KM_TYPE_NR 17
|
||||||
/* arch specific kmaps - change the numbers attached to these at your peril */
|
|
||||||
__KM_CACHE, /* cache flush page attachment point */
|
|
||||||
__KM_PGD, /* current page directory */
|
|
||||||
__KM_ITLB_PTD, /* current instruction TLB miss page table lookup */
|
|
||||||
__KM_DTLB_PTD, /* current data TLB miss page table lookup */
|
|
||||||
|
|
||||||
/* general kmaps */
|
|
||||||
KM_BOUNCE_READ,
|
|
||||||
KM_SKB_SUNRPC_DATA,
|
|
||||||
KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KM_USER0,
|
|
||||||
KM_USER1,
|
|
||||||
KM_BIO_SRC_IRQ,
|
|
||||||
KM_BIO_DST_IRQ,
|
|
||||||
KM_PTE0,
|
|
||||||
KM_PTE1,
|
|
||||||
KM_IRQ0,
|
|
||||||
KM_IRQ1,
|
|
||||||
KM_SOFTIRQ0,
|
|
||||||
KM_SOFTIRQ1,
|
|
||||||
KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -62,14 +62,14 @@ int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents,
|
||||||
dampr2 = __get_DAMPR(2);
|
dampr2 = __get_DAMPR(2);
|
||||||
|
|
||||||
for (i = 0; i < nents; i++) {
|
for (i = 0; i < nents; i++) {
|
||||||
vaddr = kmap_atomic_primary(sg_page(&sg[i]), __KM_CACHE);
|
vaddr = kmap_atomic_primary(sg_page(&sg[i]));
|
||||||
|
|
||||||
frv_dcache_writeback((unsigned long) vaddr,
|
frv_dcache_writeback((unsigned long) vaddr,
|
||||||
(unsigned long) vaddr + PAGE_SIZE);
|
(unsigned long) vaddr + PAGE_SIZE);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
kunmap_atomic_primary(vaddr, __KM_CACHE);
|
kunmap_atomic_primary(vaddr);
|
||||||
if (dampr2) {
|
if (dampr2) {
|
||||||
__set_DAMPR(2, dampr2);
|
__set_DAMPR(2, dampr2);
|
||||||
__set_IAMPR(2, dampr2);
|
__set_IAMPR(2, dampr2);
|
||||||
|
|
|
@ -26,11 +26,11 @@ void flush_dcache_page(struct page *page)
|
||||||
|
|
||||||
dampr2 = __get_DAMPR(2);
|
dampr2 = __get_DAMPR(2);
|
||||||
|
|
||||||
vaddr = kmap_atomic_primary(page, __KM_CACHE);
|
vaddr = kmap_atomic_primary(page);
|
||||||
|
|
||||||
frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE);
|
frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE);
|
||||||
|
|
||||||
kunmap_atomic_primary(vaddr, __KM_CACHE);
|
kunmap_atomic_primary(vaddr);
|
||||||
|
|
||||||
if (dampr2) {
|
if (dampr2) {
|
||||||
__set_DAMPR(2, dampr2);
|
__set_DAMPR(2, dampr2);
|
||||||
|
@ -54,12 +54,12 @@ void flush_icache_user_range(struct vm_area_struct *vma, struct page *page,
|
||||||
|
|
||||||
dampr2 = __get_DAMPR(2);
|
dampr2 = __get_DAMPR(2);
|
||||||
|
|
||||||
vaddr = kmap_atomic_primary(page, __KM_CACHE);
|
vaddr = kmap_atomic_primary(page);
|
||||||
|
|
||||||
start = (start & ~PAGE_MASK) | (unsigned long) vaddr;
|
start = (start & ~PAGE_MASK) | (unsigned long) vaddr;
|
||||||
frv_cache_wback_inv(start, start + len);
|
frv_cache_wback_inv(start, start + len);
|
||||||
|
|
||||||
kunmap_atomic_primary(vaddr, __KM_CACHE);
|
kunmap_atomic_primary(vaddr);
|
||||||
|
|
||||||
if (dampr2) {
|
if (dampr2) {
|
||||||
__set_DAMPR(2, dampr2);
|
__set_DAMPR(2, dampr2);
|
||||||
|
|
|
@ -50,11 +50,11 @@ void *kmap_atomic(struct page *page)
|
||||||
/*
|
/*
|
||||||
* The first 4 primary maps are reserved for architecture code
|
* The first 4 primary maps are reserved for architecture code
|
||||||
*/
|
*/
|
||||||
case 0: return __kmap_atomic_primary(4, paddr, 6);
|
case 0: return __kmap_atomic_primary(0, paddr, 6);
|
||||||
case 1: return __kmap_atomic_primary(5, paddr, 7);
|
case 1: return __kmap_atomic_primary(0, paddr, 7);
|
||||||
case 2: return __kmap_atomic_primary(6, paddr, 8);
|
case 2: return __kmap_atomic_primary(0, paddr, 8);
|
||||||
case 3: return __kmap_atomic_primary(7, paddr, 9);
|
case 3: return __kmap_atomic_primary(0, paddr, 9);
|
||||||
case 4: return __kmap_atomic_primary(8, paddr, 10);
|
case 4: return __kmap_atomic_primary(0, paddr, 10);
|
||||||
|
|
||||||
case 5 ... 5 + NR_TLB_LINES - 1:
|
case 5 ... 5 + NR_TLB_LINES - 1:
|
||||||
return __kmap_atomic_secondary(type - 5, paddr);
|
return __kmap_atomic_secondary(type - 5, paddr);
|
||||||
|
@ -70,11 +70,11 @@ void __kunmap_atomic(void *kvaddr)
|
||||||
{
|
{
|
||||||
int type = kmap_atomic_idx();
|
int type = kmap_atomic_idx();
|
||||||
switch (type) {
|
switch (type) {
|
||||||
case 0: __kunmap_atomic_primary(4, 6); break;
|
case 0: __kunmap_atomic_primary(0, 6); break;
|
||||||
case 1: __kunmap_atomic_primary(5, 7); break;
|
case 1: __kunmap_atomic_primary(0, 7); break;
|
||||||
case 2: __kunmap_atomic_primary(6, 8); break;
|
case 2: __kunmap_atomic_primary(0, 8); break;
|
||||||
case 3: __kunmap_atomic_primary(7, 9); break;
|
case 3: __kunmap_atomic_primary(0, 9); break;
|
||||||
case 4: __kunmap_atomic_primary(8, 10); break;
|
case 4: __kunmap_atomic_primary(0, 10); break;
|
||||||
|
|
||||||
case 5 ... 5 + NR_TLB_LINES - 1:
|
case 5 ... 5 + NR_TLB_LINES - 1:
|
||||||
__kunmap_atomic_secondary(type - 5, kvaddr);
|
__kunmap_atomic_secondary(type - 5, kvaddr);
|
||||||
|
|
|
@ -10,36 +10,7 @@
|
||||||
* 2 of the License, or (at your option) any later version.
|
* 2 of the License, or (at your option) any later version.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
enum km_type {
|
#define KM_TYPE_NR 16
|
||||||
KM_BOUNCE_READ,
|
|
||||||
KM_SKB_SUNRPC_DATA,
|
|
||||||
KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KM_USER0,
|
|
||||||
KM_USER1,
|
|
||||||
KM_BIO_SRC_IRQ,
|
|
||||||
KM_BIO_DST_IRQ,
|
|
||||||
KM_PTE0,
|
|
||||||
KM_PTE1,
|
|
||||||
KM_IRQ0,
|
|
||||||
KM_IRQ1,
|
|
||||||
KM_SOFTIRQ0,
|
|
||||||
KM_SOFTIRQ1,
|
|
||||||
KM_PPC_SYNC_PAGE,
|
|
||||||
KM_PPC_SYNC_ICACHE,
|
|
||||||
KM_KDB,
|
|
||||||
KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
/*
|
|
||||||
* This is a temporary build fix that (so they say on lkml....) should no longer
|
|
||||||
* be required after 2.6.33, because of changes planned to the kmap code.
|
|
||||||
* Let's try to remove this cruft then.
|
|
||||||
*/
|
|
||||||
#ifdef CONFIG_DEBUG_HIGHMEM
|
|
||||||
#define KM_NMI (-1)
|
|
||||||
#define KM_NMI_PTE (-1)
|
|
||||||
#define KM_IRQ_PTE (-1)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif /* __KERNEL__ */
|
#endif /* __KERNEL__ */
|
||||||
#endif /* _ASM_POWERPC_KMAP_TYPES_H */
|
#endif /* _ASM_POWERPC_KMAP_TYPES_H */
|
||||||
|
|
|
@ -23,35 +23,6 @@
|
||||||
* adds 4MB of required address-space. For now we leave KM_TYPE_NR
|
* adds 4MB of required address-space. For now we leave KM_TYPE_NR
|
||||||
* set to depth 8.
|
* set to depth 8.
|
||||||
*/
|
*/
|
||||||
enum km_type {
|
#define KM_TYPE_NR 8
|
||||||
KM_TYPE_NR = 8
|
|
||||||
};
|
|
||||||
|
|
||||||
/*
|
|
||||||
* We provide dummy definitions of all the stray values that used to be
|
|
||||||
* required for kmap_atomic() and no longer are.
|
|
||||||
*/
|
|
||||||
enum {
|
|
||||||
KM_BOUNCE_READ,
|
|
||||||
KM_SKB_SUNRPC_DATA,
|
|
||||||
KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KM_USER0,
|
|
||||||
KM_USER1,
|
|
||||||
KM_BIO_SRC_IRQ,
|
|
||||||
KM_BIO_DST_IRQ,
|
|
||||||
KM_PTE0,
|
|
||||||
KM_PTE1,
|
|
||||||
KM_IRQ0,
|
|
||||||
KM_IRQ1,
|
|
||||||
KM_SOFTIRQ0,
|
|
||||||
KM_SOFTIRQ1,
|
|
||||||
KM_SYNC_ICACHE,
|
|
||||||
KM_SYNC_DCACHE,
|
|
||||||
KM_UML_USERCOPY,
|
|
||||||
KM_IRQ_PTE,
|
|
||||||
KM_NMI,
|
|
||||||
KM_NMI_PTE,
|
|
||||||
KM_KDB
|
|
||||||
};
|
|
||||||
|
|
||||||
#endif /* _ASM_TILE_KMAP_TYPES_H */
|
#endif /* _ASM_TILE_KMAP_TYPES_H */
|
||||||
|
|
|
@ -93,7 +93,7 @@ static DEFINE_PER_CPU(struct kmap_amps, amps);
|
||||||
* If we examine it earlier we are exposed to a race where it looks
|
* If we examine it earlier we are exposed to a race where it looks
|
||||||
* writable earlier, but becomes immutable before we write the PTE.
|
* writable earlier, but becomes immutable before we write the PTE.
|
||||||
*/
|
*/
|
||||||
static void kmap_atomic_register(struct page *page, enum km_type type,
|
static void kmap_atomic_register(struct page *page, int type,
|
||||||
unsigned long va, pte_t *ptep, pte_t pteval)
|
unsigned long va, pte_t *ptep, pte_t pteval)
|
||||||
{
|
{
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
|
@ -8,22 +8,6 @@
|
||||||
|
|
||||||
/* No more #include "asm/arch/kmap_types.h" ! */
|
/* No more #include "asm/arch/kmap_types.h" ! */
|
||||||
|
|
||||||
enum km_type {
|
#define KM_TYPE_NR 14
|
||||||
KM_BOUNCE_READ,
|
|
||||||
KM_SKB_SUNRPC_DATA,
|
|
||||||
KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KM_USER0,
|
|
||||||
KM_USER1,
|
|
||||||
KM_UML_USERCOPY, /* UML specific, for copy_*_user - used in do_op_one_page */
|
|
||||||
KM_BIO_SRC_IRQ,
|
|
||||||
KM_BIO_DST_IRQ,
|
|
||||||
KM_PTE0,
|
|
||||||
KM_PTE1,
|
|
||||||
KM_IRQ0,
|
|
||||||
KM_IRQ1,
|
|
||||||
KM_SOFTIRQ0,
|
|
||||||
KM_SOFTIRQ1,
|
|
||||||
KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -349,12 +349,12 @@ static void jbd2_block_tag_csum_set(journal_t *j, journal_block_tag_t *tag,
|
||||||
return;
|
return;
|
||||||
|
|
||||||
sequence = cpu_to_be32(sequence);
|
sequence = cpu_to_be32(sequence);
|
||||||
addr = kmap_atomic(page, KM_USER0);
|
addr = kmap_atomic(page);
|
||||||
csum = jbd2_chksum(j, j->j_csum_seed, (__u8 *)&sequence,
|
csum = jbd2_chksum(j, j->j_csum_seed, (__u8 *)&sequence,
|
||||||
sizeof(sequence));
|
sizeof(sequence));
|
||||||
csum = jbd2_chksum(j, csum, addr + offset_in_page(bh->b_data),
|
csum = jbd2_chksum(j, csum, addr + offset_in_page(bh->b_data),
|
||||||
bh->b_size);
|
bh->b_size);
|
||||||
kunmap_atomic(addr, KM_USER0);
|
kunmap_atomic(addr);
|
||||||
|
|
||||||
tag->t_checksum = cpu_to_be32(csum);
|
tag->t_checksum = cpu_to_be32(csum);
|
||||||
}
|
}
|
||||||
|
|
|
@ -224,7 +224,7 @@ static void anon_pipe_buf_release(struct pipe_inode_info *pipe,
|
||||||
* and the caller has to be careful not to fault before calling
|
* and the caller has to be careful not to fault before calling
|
||||||
* the unmap function.
|
* the unmap function.
|
||||||
*
|
*
|
||||||
* Note that this function occupies KM_USER0 if @atomic != 0.
|
* Note that this function calls kmap_atomic() if @atomic != 0.
|
||||||
*/
|
*/
|
||||||
void *generic_pipe_buf_map(struct pipe_inode_info *pipe,
|
void *generic_pipe_buf_map(struct pipe_inode_info *pipe,
|
||||||
struct pipe_buffer *buf, int atomic)
|
struct pipe_buffer *buf, int atomic)
|
||||||
|
|
|
@ -2,39 +2,9 @@
|
||||||
#define _ASM_GENERIC_KMAP_TYPES_H
|
#define _ASM_GENERIC_KMAP_TYPES_H
|
||||||
|
|
||||||
#ifdef __WITH_KM_FENCE
|
#ifdef __WITH_KM_FENCE
|
||||||
# define KMAP_D(n) __KM_FENCE_##n ,
|
# define KM_TYPE_NR 41
|
||||||
#else
|
#else
|
||||||
# define KMAP_D(n)
|
# define KM_TYPE_NR 20
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
enum km_type {
|
|
||||||
KMAP_D(0) KM_BOUNCE_READ,
|
|
||||||
KMAP_D(1) KM_SKB_SUNRPC_DATA,
|
|
||||||
KMAP_D(2) KM_SKB_DATA_SOFTIRQ,
|
|
||||||
KMAP_D(3) KM_USER0,
|
|
||||||
KMAP_D(4) KM_USER1,
|
|
||||||
KMAP_D(5) KM_BIO_SRC_IRQ,
|
|
||||||
KMAP_D(6) KM_BIO_DST_IRQ,
|
|
||||||
KMAP_D(7) KM_PTE0,
|
|
||||||
KMAP_D(8) KM_PTE1,
|
|
||||||
KMAP_D(9) KM_IRQ0,
|
|
||||||
KMAP_D(10) KM_IRQ1,
|
|
||||||
KMAP_D(11) KM_SOFTIRQ0,
|
|
||||||
KMAP_D(12) KM_SOFTIRQ1,
|
|
||||||
KMAP_D(13) KM_SYNC_ICACHE,
|
|
||||||
KMAP_D(14) KM_SYNC_DCACHE,
|
|
||||||
/* UML specific, for copy_*_user - used in do_op_one_page */
|
|
||||||
KMAP_D(15) KM_UML_USERCOPY,
|
|
||||||
KMAP_D(16) KM_IRQ_PTE,
|
|
||||||
KMAP_D(17) KM_NMI,
|
|
||||||
KMAP_D(18) KM_NMI_PTE,
|
|
||||||
KMAP_D(19) KM_KDB,
|
|
||||||
/*
|
|
||||||
* Remember to update debug_kmap_atomic() when adding new kmap types!
|
|
||||||
*/
|
|
||||||
KMAP_D(20) KM_TYPE_NR
|
|
||||||
};
|
|
||||||
|
|
||||||
#undef KMAP_D
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -109,55 +109,16 @@ static inline void kmap_atomic_idx_pop(void)
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/*
|
|
||||||
* NOTE:
|
|
||||||
* kmap_atomic() and kunmap_atomic() with two arguments are deprecated.
|
|
||||||
* We only keep them for backward compatibility, any usage of them
|
|
||||||
* are now warned.
|
|
||||||
*/
|
|
||||||
|
|
||||||
#define PASTE(a, b) a ## b
|
|
||||||
#define PASTE2(a, b) PASTE(a, b)
|
|
||||||
|
|
||||||
#define NARG_(_2, _1, n, ...) n
|
|
||||||
#define NARG(...) NARG_(__VA_ARGS__, 2, 1, :)
|
|
||||||
|
|
||||||
static inline void __deprecated *kmap_atomic_deprecated(struct page *page,
|
|
||||||
enum km_type km)
|
|
||||||
{
|
|
||||||
return kmap_atomic(page);
|
|
||||||
}
|
|
||||||
|
|
||||||
#define kmap_atomic1(...) kmap_atomic(__VA_ARGS__)
|
|
||||||
#define kmap_atomic2(...) kmap_atomic_deprecated(__VA_ARGS__)
|
|
||||||
#define kmap_atomic(...) PASTE2(kmap_atomic, NARG(__VA_ARGS__)(__VA_ARGS__))
|
|
||||||
|
|
||||||
static inline void __deprecated __kunmap_atomic_deprecated(void *addr,
|
|
||||||
enum km_type km)
|
|
||||||
{
|
|
||||||
__kunmap_atomic(addr);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Prevent people trying to call kunmap_atomic() as if it were kunmap()
|
* Prevent people trying to call kunmap_atomic() as if it were kunmap()
|
||||||
* kunmap_atomic() should get the return value of kmap_atomic, not the page.
|
* kunmap_atomic() should get the return value of kmap_atomic, not the page.
|
||||||
*/
|
*/
|
||||||
#define kunmap_atomic_deprecated(addr, km) \
|
#define kunmap_atomic(addr) \
|
||||||
do { \
|
|
||||||
BUILD_BUG_ON(__same_type((addr), struct page *)); \
|
|
||||||
__kunmap_atomic_deprecated(addr, km); \
|
|
||||||
} while (0)
|
|
||||||
|
|
||||||
#define kunmap_atomic_withcheck(addr) \
|
|
||||||
do { \
|
do { \
|
||||||
BUILD_BUG_ON(__same_type((addr), struct page *)); \
|
BUILD_BUG_ON(__same_type((addr), struct page *)); \
|
||||||
__kunmap_atomic(addr); \
|
__kunmap_atomic(addr); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
#define kunmap_atomic1(...) kunmap_atomic_withcheck(__VA_ARGS__)
|
|
||||||
#define kunmap_atomic2(...) kunmap_atomic_deprecated(__VA_ARGS__)
|
|
||||||
#define kunmap_atomic(...) PASTE2(kunmap_atomic, NARG(__VA_ARGS__)(__VA_ARGS__))
|
|
||||||
/**** End of C pre-processor tricks for deprecated macros ****/
|
|
||||||
|
|
||||||
/* when CONFIG_HIGHMEM is not set these will be plain clear/copy_page */
|
/* when CONFIG_HIGHMEM is not set these will be plain clear/copy_page */
|
||||||
#ifndef clear_user_highpage
|
#ifndef clear_user_highpage
|
||||||
|
|
|
@ -86,11 +86,9 @@ struct pipe_buf_operations {
|
||||||
* mapping or not. The atomic map is faster, however you can't take
|
* mapping or not. The atomic map is faster, however you can't take
|
||||||
* page faults before calling ->unmap() again. So if you need to eg
|
* page faults before calling ->unmap() again. So if you need to eg
|
||||||
* access user data through copy_to/from_user(), then you must get
|
* access user data through copy_to/from_user(), then you must get
|
||||||
* a non-atomic map. ->map() uses the KM_USER0 atomic slot for
|
* a non-atomic map. ->map() uses the kmap_atomic slot for
|
||||||
* atomic maps, so you can't map more than one pipe_buffer at once
|
* atomic maps, you have to be careful if mapping another page as
|
||||||
* and you have to be careful if mapping another page as source
|
* source or destination for a copy.
|
||||||
* or destination for a copy (IOW, it has to use something else
|
|
||||||
* than KM_USER0).
|
|
||||||
*/
|
*/
|
||||||
void * (*map)(struct pipe_inode_info *, struct pipe_buffer *, int);
|
void * (*map)(struct pipe_inode_info *, struct pipe_buffer *, int);
|
||||||
|
|
||||||
|
|
|
@ -1975,9 +1975,7 @@ static int aligned_vwrite(char *buf, char *addr, unsigned long count)
|
||||||
* IOREMAP area is treated as memory hole and no copy is done.
|
* IOREMAP area is treated as memory hole and no copy is done.
|
||||||
*
|
*
|
||||||
* If [addr...addr+count) doesn't includes any intersects with alive
|
* If [addr...addr+count) doesn't includes any intersects with alive
|
||||||
* vm_struct area, returns 0.
|
* vm_struct area, returns 0. @buf should be kernel's buffer.
|
||||||
* @buf should be kernel's buffer. Because this function uses KM_USER0,
|
|
||||||
* the caller should guarantee KM_USER0 is not used.
|
|
||||||
*
|
*
|
||||||
* Note: In usual ops, vread() is never necessary because the caller
|
* Note: In usual ops, vread() is never necessary because the caller
|
||||||
* should know vmalloc() area is valid and can use memcpy().
|
* should know vmalloc() area is valid and can use memcpy().
|
||||||
|
@ -2051,9 +2049,7 @@ finished:
|
||||||
* IOREMAP area is treated as memory hole and no copy is done.
|
* IOREMAP area is treated as memory hole and no copy is done.
|
||||||
*
|
*
|
||||||
* If [addr...addr+count) doesn't includes any intersects with alive
|
* If [addr...addr+count) doesn't includes any intersects with alive
|
||||||
* vm_struct area, returns 0.
|
* vm_struct area, returns 0. @buf should be kernel's buffer.
|
||||||
* @buf should be kernel's buffer. Because this function uses KM_USER0,
|
|
||||||
* the caller should guarantee KM_USER0 is not used.
|
|
||||||
*
|
*
|
||||||
* Note: In usual ops, vwrite() is never necessary because the caller
|
* Note: In usual ops, vwrite() is never necessary because the caller
|
||||||
* should know vmalloc() area is valid and can use memcpy().
|
* should know vmalloc() area is valid and can use memcpy().
|
||||||
|
|
Loading…
Add table
Reference in a new issue