Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
1b3bf847
Commit
1b3bf847
authored
Mar 04, 2016
by
Russell King
Browse files
Options
Browse Files
Download
Plain Diff
Merge branches 'amba', 'fixes', 'misc' and 'tauros2' into for-next
parents
18558cae
656b8035
f474c8c8
b4268676
1d93ba2a
Changes
59
Hide whitespace changes
Inline
Side-by-side
Showing
59 changed files
with
975 additions
and
377 deletions
+975
-377
Documentation/DMA-attributes.txt
Documentation/DMA-attributes.txt
+26
-0
Documentation/driver-model/porting.txt
Documentation/driver-model/porting.txt
+4
-2
arch/arm/Kconfig
arch/arm/Kconfig
+2
-3
arch/arm/Makefile
arch/arm/Makefile
+0
-1
arch/arm/boot/Makefile
arch/arm/boot/Makefile
+1
-9
arch/arm/boot/compressed/.gitignore
arch/arm/boot/compressed/.gitignore
+1
-5
arch/arm/boot/compressed/Makefile
arch/arm/boot/compressed/Makefile
+16
-17
arch/arm/boot/compressed/piggy.S
arch/arm/boot/compressed/piggy.S
+1
-1
arch/arm/boot/compressed/piggy.lz4.S
arch/arm/boot/compressed/piggy.lz4.S
+0
-6
arch/arm/boot/compressed/piggy.lzma.S
arch/arm/boot/compressed/piggy.lzma.S
+0
-6
arch/arm/boot/compressed/piggy.lzo.S
arch/arm/boot/compressed/piggy.lzo.S
+0
-6
arch/arm/boot/compressed/piggy.xzkern.S
arch/arm/boot/compressed/piggy.xzkern.S
+0
-6
arch/arm/common/sa1111.c
arch/arm/common/sa1111.c
+1
-1
arch/arm/include/asm/Kbuild
arch/arm/include/asm/Kbuild
+0
-1
arch/arm/include/asm/div64.h
arch/arm/include/asm/div64.h
+7
-7
arch/arm/include/asm/memory.h
arch/arm/include/asm/memory.h
+18
-17
arch/arm/include/asm/mmu_context.h
arch/arm/include/asm/mmu_context.h
+12
-2
arch/arm/include/asm/sections.h
arch/arm/include/asm/sections.h
+8
-0
arch/arm/include/asm/sparsemem.h
arch/arm/include/asm/sparsemem.h
+4
-3
arch/arm/kernel/Makefile
arch/arm/kernel/Makefile
+1
-0
arch/arm/kernel/entry-armv.S
arch/arm/kernel/entry-armv.S
+3
-5
arch/arm/kernel/hibernate.c
arch/arm/kernel/hibernate.c
+2
-2
arch/arm/kernel/hyp-stub.S
arch/arm/kernel/hyp-stub.S
+24
-0
arch/arm/kernel/irq.c
arch/arm/kernel/irq.c
+1
-1
arch/arm/kernel/machine_kexec.c
arch/arm/kernel/machine_kexec.c
+6
-10
arch/arm/kernel/module.c
arch/arm/kernel/module.c
+1
-1
arch/arm/kernel/reboot.c
arch/arm/kernel/reboot.c
+1
-1
arch/arm/kernel/topology.c
arch/arm/kernel/topology.c
+1
-3
arch/arm/kernel/vmlinux-xip.lds.S
arch/arm/kernel/vmlinux-xip.lds.S
+316
-0
arch/arm/kernel/vmlinux.lds.S
arch/arm/kernel/vmlinux.lds.S
+26
-34
arch/arm/mach-davinci/include/mach/uncompress.h
arch/arm/mach-davinci/include/mach/uncompress.h
+1
-1
arch/arm/mach-footbridge/Kconfig
arch/arm/mach-footbridge/Kconfig
+0
-1
arch/arm/mach-keystone/keystone.c
arch/arm/mach-keystone/keystone.c
+1
-1
arch/arm/mach-ks8695/include/mach/uncompress.h
arch/arm/mach-ks8695/include/mach/uncompress.h
+1
-1
arch/arm/mach-netx/include/mach/uncompress.h
arch/arm/mach-netx/include/mach/uncompress.h
+1
-1
arch/arm/mach-omap1/include/mach/uncompress.h
arch/arm/mach-omap1/include/mach/uncompress.h
+1
-1
arch/arm/mach-rpc/include/mach/uncompress.h
arch/arm/mach-rpc/include/mach/uncompress.h
+1
-1
arch/arm/mach-sa1100/include/mach/uncompress.h
arch/arm/mach-sa1100/include/mach/uncompress.h
+1
-1
arch/arm/mach-w90x900/include/mach/uncompress.h
arch/arm/mach-w90x900/include/mach/uncompress.h
+1
-1
arch/arm/mm/Kconfig
arch/arm/mm/Kconfig
+19
-17
arch/arm/mm/cache-tauros2.c
arch/arm/mm/cache-tauros2.c
+18
-14
arch/arm/mm/dma-mapping.c
arch/arm/mm/dma-mapping.c
+198
-50
arch/arm/mm/idmap.c
arch/arm/mm/idmap.c
+1
-1
arch/arm/mm/init.c
arch/arm/mm/init.c
+13
-11
arch/arm/mm/mmu.c
arch/arm/mm/mmu.c
+6
-2
arch/arm/mm/pageattr.c
arch/arm/mm/pageattr.c
+3
-0
arch/arm/mm/proc-v7.S
arch/arm/mm/proc-v7.S
+1
-1
arch/arm/plat-orion/time.c
arch/arm/plat-orion/time.c
+13
-0
arch/arm64/kernel/psci.c
arch/arm64/kernel/psci.c
+2
-97
drivers/base/dd.c
drivers/base/dd.c
+22
-2
drivers/clk/clkdev.c
drivers/clk/clkdev.c
+25
-6
drivers/firmware/psci.c
drivers/firmware/psci.c
+120
-0
drivers/media/v4l2-core/videobuf2-dma-contig.c
drivers/media/v4l2-core/videobuf2-dma-contig.c
+22
-11
drivers/nvdimm/bus.c
drivers/nvdimm/bus.c
+1
-1
include/linux/clkdev.h
include/linux/clkdev.h
+1
-2
include/linux/device.h
include/linux/device.h
+5
-2
include/linux/dma-attrs.h
include/linux/dma-attrs.h
+1
-0
include/linux/psci.h
include/linux/psci.h
+3
-0
include/media/videobuf2-dma-contig.h
include/media/videobuf2-dma-contig.h
+10
-1
No files found.
Documentation/DMA-attributes.txt
View file @
1b3bf847
...
...
@@ -100,3 +100,29 @@ allocated by dma_alloc_attrs() function from individual pages if it can
be mapped as contiguous chunk into device dma address space. By
specifying this attribute the allocated buffer is forced to be contiguous
also in physical memory.
DMA_ATTR_ALLOC_SINGLE_PAGES
---------------------------
This is a hint to the DMA-mapping subsystem that it's probably not worth
the time to try to allocate memory to in a way that gives better TLB
efficiency (AKA it's not worth trying to build the mapping out of larger
pages). You might want to specify this if:
- You know that the accesses to this memory won't thrash the TLB.
You might know that the accesses are likely to be sequential or
that they aren't sequential but it's unlikely you'll ping-pong
between many addresses that are likely to be in different physical
pages.
- You know that the penalty of TLB misses while accessing the
memory will be small enough to be inconsequential. If you are
doing a heavy operation like decryption or decompression this
might be the case.
- You know that the DMA mapping is fairly transitory. If you expect
the mapping to have a short lifetime then it may be worth it to
optimize allocation (avoid coming up with large pages) instead of
getting the slight performance win of larger pages.
Setting this hint doesn't guarantee that you won't get huge pages, but it
means that we won't try quite as hard to get them.
NOTE: At the moment DMA_ATTR_ALLOC_SINGLE_PAGES is only implemented on ARM,
though ARM64 patches will likely be posted soon.
Documentation/driver-model/porting.txt
View file @
1b3bf847
...
...
@@ -340,8 +340,10 @@ comparison:
int (*match)(struct device * dev, struct device_driver * drv);
match should return '1' if the driver supports the device, and '0'
otherwise.
match should return positive value if the driver supports the device,
and zero otherwise. It may also return error code (for example
-EPROBE_DEFER) if determining that given driver supports the device is
not possible.
When a device is registered, the bus's list of drivers is iterated
over. bus->match() is called for each one until a match is found.
...
...
arch/arm/Kconfig
View file @
1b3bf847
...
...
@@ -572,7 +572,6 @@ config ARCH_RPC
select NEED_MACH_IO_H
select NEED_MACH_MEMORY_H
select NO_IOPORT_MAP
select VIRT_TO_BUS
help
On the Acorn Risc-PC, Linux can support the internal IDE disk and
CD-ROM interface, serial and parallel port, and the floppy drive.
...
...
@@ -1337,7 +1336,6 @@ config BIG_LITTLE
config BL_SWITCHER
bool "big.LITTLE switcher support"
depends on BIG_LITTLE && MCPM && HOTPLUG_CPU && ARM_GIC
select ARM_CPU_SUSPEND
select CPU_PM
help
The big.LITTLE "switcher" provides the core functionality to
...
...
@@ -2111,7 +2109,8 @@ config ARCH_SUSPEND_POSSIBLE
def_bool y
config ARM_CPU_SUSPEND
def_bool PM_SLEEP
def_bool PM_SLEEP || BL_SWITCHER || ARM_PSCI_FW
depends on ARCH_SUSPEND_POSSIBLE
config ARCH_HIBERNATION_POSSIBLE
bool
...
...
arch/arm/Makefile
View file @
1b3bf847
...
...
@@ -352,7 +352,6 @@ archclean:
# My testing targets (bypasses dependencies)
bp
:
; $(Q)$(MAKE) $(build)=$(boot) MACHINE=$(MACHINE) $(boot)/bootpImage
i zi
:
; $(Q)$(MAKE) $(build)=$(boot) MACHINE=$(MACHINE) $@
define
archhelp
...
...
arch/arm/boot/Makefile
View file @
1b3bf847
...
...
@@ -88,7 +88,7 @@ $(obj)/bootpImage: $(obj)/bootp/bootp FORCE
$(
call
if_changed,objcopy
)
@
$(kecho)
' Kernel: $@ is ready'
PHONY
+=
initrd
FORCE
PHONY
+=
initrd
initrd
:
@
test
"
$(INITRD_PHYS)
"
!=
""
||
\
(
echo
This machine does not support INITRD
;
exit
-1
)
...
...
@@ -107,12 +107,4 @@ uinstall:
$(CONFIG_SHELL)
$(srctree)
/
$(src)
/install.sh
"
$(KERNELRELEASE)
"
\
$(obj)
/uImage System.map
"
$(INSTALL_PATH)
"
zi
:
$(CONFIG_SHELL)
$(srctree)
/
$(src)
/install.sh
"
$(KERNELRELEASE)
"
\
$(obj)
/zImage System.map
"
$(INSTALL_PATH)
"
i
:
$(CONFIG_SHELL)
$(srctree)
/
$(src)
/install.sh
"
$(KERNELRELEASE)
"
\
$(obj)
/Image System.map
"
$(INSTALL_PATH)
"
subdir-
:=
bootp compressed dts
arch/arm/boot/compressed/.gitignore
View file @
1b3bf847
...
...
@@ -3,11 +3,7 @@ bswapsdi2.S
font.c
lib1funcs.S
hyp-stub.S
piggy.gzip
piggy.lzo
piggy.lzma
piggy.xzkern
piggy.lz4
piggy_data
vmlinux
vmlinux.lds
...
...
arch/arm/boot/compressed/Makefile
View file @
1b3bf847
...
...
@@ -66,11 +66,11 @@ endif
CPPFLAGS_vmlinux.lds
:=
-DTEXT_START
=
"
$(ZTEXTADDR)
"
-DBSS_START
=
"
$(ZBSSADDR)
"
suffix_
$(CONFIG_KERNEL_GZIP)
=
gzip
suffix_
$(CONFIG_KERNEL_LZO)
=
lzo
suffix_
$(CONFIG_KERNEL_LZMA)
=
lzma
suffix_
$(CONFIG_KERNEL_XZ)
=
xzkern
suffix_
$(CONFIG_KERNEL_LZ4)
=
lz4
compress-
$(CONFIG_KERNEL_GZIP)
=
gzip
compress-
$(CONFIG_KERNEL_LZO)
=
lzo
compress-
$(CONFIG_KERNEL_LZMA)
=
lzma
compress-
$(CONFIG_KERNEL_XZ)
=
xzkern
compress-
$(CONFIG_KERNEL_LZ4)
=
lz4
# Borrowed libfdt files for the ATAG compatibility mode
...
...
@@ -89,15 +89,12 @@ ifeq ($(CONFIG_ARM_ATAG_DTB_COMPAT),y)
OBJS
+=
$(libfdt_objs)
atags_to_fdt.o
endif
targets
:=
vmlinux vmlinux.lds
\
piggy.
$(suffix_y)
piggy.
$(suffix_y)
.o
\
lib1funcs.o lib1funcs.S ashldi3.o ashldi3.S bswapsdi2.o
\
bswapsdi2.S font.o font.c head.o misc.o
$(OBJS)
targets
:=
vmlinux vmlinux.lds piggy_data piggy.o
\
lib1funcs.o ashldi3.o bswapsdi2.o
\
head.o
$(OBJS)
# Make sure files are removed during clean
extra-y
+=
piggy.gzip piggy.lzo piggy.lzma piggy.xzkern piggy.lz4
\
lib1funcs.S ashldi3.S bswapsdi2.S
$(libfdt)
$(libfdt_hdrs)
\
hyp-stub.S
clean-files
+=
piggy_data lib1funcs.S ashldi3.S bswapsdi2.S
\
$(libfdt)
$(libfdt_hdrs)
hyp-stub.S
KBUILD_CFLAGS
+=
-DDISABLE_BRANCH_PROFILING
...
...
@@ -178,22 +175,24 @@ fi
efi-obj-$(CONFIG_EFI_STUB)
:=
$(objtree)
/drivers/firmware/efi/libstub/lib.a
$(obj)/vmlinux
:
$(obj)/vmlinux.lds $(obj)/$(HEAD) $(obj)/piggy.
$(suffix_y).
o
\
$(obj)/vmlinux
:
$(obj)/vmlinux.lds $(obj)/$(HEAD) $(obj)/piggy.o
\
$(addprefix $(obj)/
,
$(OBJS)) $(lib1funcs) $(ashldi3)
\
$(bswapsdi2) $(efi-obj-y) FORCE
@
$(check_for_multiple_zreladdr)
$(
call
if_changed,ld
)
@
$(check_for_bad_syms)
$(obj)/piggy
.$(suffix_y)
:
$(obj)/../Image FORCE
$(
call
if_changed,
$(
suffix_
y)
)
$(obj)/piggy
_data
:
$(obj)/../Image FORCE
$(
call
if_changed,
$
(
compress-
y
))
$(obj)/piggy.
$(suffix_y).o
:
$(obj)/piggy.$(suffix_y) FORCE
$(obj)/piggy.
o
:
$(obj)/piggy_data
CFLAGS_font.o
:=
-Dstatic
=
$(obj)/font.c
:
$(FONTC)
$(
call
cmd,shipped
)
AFLAGS_hyp-stub.o
:=
-Wa
,-march
=
armv7-a
$(obj)/hyp-stub.S
:
$(srctree)/arch/$(SRCARCH)/kernel/hyp-stub.S
$(
call
cmd,shipped
)
arch/arm/boot/compressed/piggy.
gzip.
S
→
arch/arm/boot/compressed/piggy.S
View file @
1b3bf847
.
section
.
piggydata
,#
alloc
.
globl
input_data
input_data
:
.
incbin
"arch/arm/boot/compressed/piggy
.gzip
"
.
incbin
"arch/arm/boot/compressed/piggy
_data
"
.
globl
input_data_end
input_data_end
:
arch/arm/boot/compressed/piggy.lz4.S
deleted
100644 → 0
View file @
18558cae
.
section
.
piggydata
,#
alloc
.
globl
input_data
input_data
:
.
incbin
"arch/arm/boot/compressed/piggy.lz4"
.
globl
input_data_end
input_data_end
:
arch/arm/boot/compressed/piggy.lzma.S
deleted
100644 → 0
View file @
18558cae
.
section
.
piggydata
,#
alloc
.
globl
input_data
input_data
:
.
incbin
"arch/arm/boot/compressed/piggy.lzma"
.
globl
input_data_end
input_data_end
:
arch/arm/boot/compressed/piggy.lzo.S
deleted
100644 → 0
View file @
18558cae
.
section
.
piggydata
,#
alloc
.
globl
input_data
input_data
:
.
incbin
"arch/arm/boot/compressed/piggy.lzo"
.
globl
input_data_end
input_data_end
:
arch/arm/boot/compressed/piggy.xzkern.S
deleted
100644 → 0
View file @
18558cae
.
section
.
piggydata
,#
alloc
.
globl
input_data
input_data
:
.
incbin
"arch/arm/boot/compressed/piggy.xzkern"
.
globl
input_data_end
input_data_end
:
arch/arm/common/sa1111.c
View file @
1b3bf847
...
...
@@ -1290,7 +1290,7 @@ static int sa1111_match(struct device *_dev, struct device_driver *_drv)
struct
sa1111_dev
*
dev
=
SA1111_DEV
(
_dev
);
struct
sa1111_driver
*
drv
=
SA1111_DRV
(
_drv
);
return
dev
->
devid
&
drv
->
devid
;
return
!!
(
dev
->
devid
&
drv
->
devid
)
;
}
static
int
sa1111_bus_suspend
(
struct
device
*
dev
,
pm_message_t
state
)
...
...
arch/arm/include/asm/Kbuild
View file @
1b3bf847
...
...
@@ -23,7 +23,6 @@ generic-y += preempt.h
generic-y += resource.h
generic-y += rwsem.h
generic-y += seccomp.h
generic-y += sections.h
generic-y += segment.h
generic-y += sembuf.h
generic-y += serial.h
...
...
arch/arm/include/asm/div64.h
View file @
1b3bf847
...
...
@@ -74,7 +74,7 @@ static inline uint32_t __div64_32(uint64_t *n, uint32_t base)
static
inline
uint64_t
__arch_xprod_64
(
uint64_t
m
,
uint64_t
n
,
bool
bias
)
{
unsigned
long
long
res
;
unsigned
int
tmp
=
0
;
register
unsigned
int
tmp
asm
(
"ip"
)
=
0
;
if
(
!
bias
)
{
asm
(
"umull %Q0, %R0, %Q1, %Q2
\n\t
"
...
...
@@ -90,12 +90,12 @@ static inline uint64_t __arch_xprod_64(uint64_t m, uint64_t n, bool bias)
:
"r"
(
m
),
"r"
(
n
)
:
"cc"
);
}
else
{
asm
(
"umull %Q0, %R0, %Q
1, %Q2
\n\t
"
"cmn %Q0, %Q
1
\n\t
"
"adcs %R0, %R0, %R
1
\n\t
"
"adc %Q0, %
3
, #0"
:
"=&r"
(
res
)
:
"r"
(
m
),
"r"
(
n
)
,
"r"
(
tmp
)
asm
(
"umull %Q0, %R0, %Q
2, %Q3
\n\t
"
"cmn %Q0, %Q
2
\n\t
"
"adcs %R0, %R0, %R
2
\n\t
"
"adc %Q0, %
1
, #0"
:
"=&r"
(
res
)
,
"+&r"
(
tmp
)
:
"r"
(
m
),
"r"
(
n
)
:
"cc"
);
}
...
...
arch/arm/include/asm/memory.h
View file @
1b3bf847
...
...
@@ -134,6 +134,21 @@
*/
#define PLAT_PHYS_OFFSET UL(CONFIG_PHYS_OFFSET)
#ifdef CONFIG_XIP_KERNEL
/*
* When referencing data in RAM from the XIP region in a relative manner
* with the MMU off, we need the relative offset between the two physical
* addresses. The macro below achieves this, which is:
* __pa(v_data) - __xip_pa(v_text)
*/
#define PHYS_RELATIVE(v_data, v_text) \
(((v_data) - PAGE_OFFSET + PLAT_PHYS_OFFSET) - \
((v_text) - XIP_VIRT_ADDR(CONFIG_XIP_PHYS_ADDR) + \
CONFIG_XIP_PHYS_ADDR))
#else
#define PHYS_RELATIVE(v_data, v_text) ((v_data) - (v_text))
#endif
#ifndef __ASSEMBLY__
/*
...
...
@@ -273,14 +288,14 @@ static inline void *phys_to_virt(phys_addr_t x)
#define __va(x) ((void *)__phys_to_virt((phys_addr_t)(x)))
#define pfn_to_kaddr(pfn) __va((phys_addr_t)(pfn) << PAGE_SHIFT)
extern
phys_addr_t
(
*
arch_virt_to_idmap
)(
unsigned
long
x
);
extern
unsigned
long
(
*
arch_virt_to_idmap
)(
unsigned
long
x
);
/*
* These are for systems that have a hardware interconnect supported alias of
* physical memory for idmap purposes. Most cases should leave these
* untouched.
* untouched.
Note: this can only return addresses less than 4GiB.
*/
static
inline
phys_addr_t
__virt_to_idmap
(
unsigned
long
x
)
static
inline
unsigned
long
__virt_to_idmap
(
unsigned
long
x
)
{
if
(
IS_ENABLED
(
CONFIG_MMU
)
&&
arch_virt_to_idmap
)
return
arch_virt_to_idmap
(
x
);
...
...
@@ -303,20 +318,6 @@ static inline phys_addr_t __virt_to_idmap(unsigned long x)
#define __bus_to_pfn(x) __phys_to_pfn(x)
#endif
#ifdef CONFIG_VIRT_TO_BUS
#define virt_to_bus virt_to_bus
static
inline
__deprecated
unsigned
long
virt_to_bus
(
void
*
x
)
{
return
__virt_to_bus
((
unsigned
long
)
x
);
}
#define bus_to_virt bus_to_virt
static
inline
__deprecated
void
*
bus_to_virt
(
unsigned
long
x
)
{
return
(
void
*
)
__bus_to_virt
(
x
);
}
#endif
/*
* Conversion between a struct page and a physical address.
*
...
...
arch/arm/include/asm/mmu_context.h
View file @
1b3bf847
...
...
@@ -26,7 +26,12 @@ void __check_vmalloc_seq(struct mm_struct *mm);
#ifdef CONFIG_CPU_HAS_ASID
void
check_and_switch_context
(
struct
mm_struct
*
mm
,
struct
task_struct
*
tsk
);
#define init_new_context(tsk,mm) ({ atomic64_set(&(mm)->context.id, 0); 0; })
static
inline
int
init_new_context
(
struct
task_struct
*
tsk
,
struct
mm_struct
*
mm
)
{
atomic64_set
(
&
mm
->
context
.
id
,
0
);
return
0
;
}
#ifdef CONFIG_ARM_ERRATA_798181
void
a15_erratum_get_cpumask
(
int
this_cpu
,
struct
mm_struct
*
mm
,
...
...
@@ -85,7 +90,12 @@ static inline void finish_arch_post_lock_switch(void)
#endif
/* CONFIG_MMU */
#define init_new_context(tsk,mm) 0
static
inline
int
init_new_context
(
struct
task_struct
*
tsk
,
struct
mm_struct
*
mm
)
{
return
0
;
}
#endif
/* CONFIG_CPU_HAS_ASID */
...
...
arch/arm/include/asm/sections.h
0 → 100644
View file @
1b3bf847
#ifndef _ASM_ARM_SECTIONS_H
#define _ASM_ARM_SECTIONS_H
#include <asm-generic/sections.h>
extern
char
_exiprom
[];
#endif
/* _ASM_ARM_SECTIONS_H */
arch/arm/include/asm/sparsemem.h
View file @
1b3bf847
...
...
@@ -15,10 +15,11 @@
* Eg, if you have 2 banks of up to 64MB at 0x80000000, 0x84000000,
* then MAX_PHYSMEM_BITS is 32, SECTION_SIZE_BITS is 26.
*
*
Define these
in your mach/memory.h.
*
These can be overridden
in your mach/memory.h.
*/
#if !defined(SECTION_SIZE_BITS) || !defined(MAX_PHYSMEM_BITS)
#error Sparsemem is not supported on this platform
#if !defined(MAX_PHYSMEM_BITS) || !defined(SECTION_SIZE_BITS)
#define MAX_PHYSMEM_BITS 36
#define SECTION_SIZE_BITS 28
#endif
#endif
arch/arm/kernel/Makefile
View file @
1b3bf847
...
...
@@ -88,6 +88,7 @@ obj-$(CONFIG_DEBUG_LL) += debug.o
obj-$(CONFIG_EARLY_PRINTK)
+=
early_printk.o
obj-$(CONFIG_ARM_VIRT_EXT)
+=
hyp-stub.o
AFLAGS_hyp-stub.o
:=
-Wa
,-march
=
armv7-a
ifeq
($(CONFIG_ARM_PSCI),y)
obj-$(CONFIG_SMP)
+=
psci_smp.o
endif
...
...
arch/arm/kernel/entry-armv.S
View file @
1b3bf847
...
...
@@ -1064,7 +1064,6 @@ ENDPROC(vector_\name)
.
endm
.
section
.
stubs
,
"ax"
,
%
progbits
__stubs_start
:
@
This
must
be
the
first
word
.
word
vector_swi
...
...
@@ -1202,14 +1201,13 @@ vector_addrexcptn:
.
long
__fiq_svc
@
e
.
long
__fiq_svc
@
f
.
globl
vector_fiq_offset
.
equ
vector_fiq_offset
,
vector_fiq
.
globl
vector_fiq
.
section
.
vectors
,
"ax"
,
%
progbits
__vectors_start
:
.
L
__vectors_start
:
W
(
b
)
vector_rst
W
(
b
)
vector_und
W
(
ldr
)
pc
,
__vectors_start
+
0x1000
W
(
ldr
)
pc
,
.
L
__vectors_start
+
0x1000
W
(
b
)
vector_pabt
W
(
b
)
vector_dabt
W
(
b
)
vector_addrexcptn
...
...
arch/arm/kernel/hibernate.c
View file @
1b3bf847
...
...
@@ -62,7 +62,7 @@ static int notrace arch_save_image(unsigned long unused)
ret
=
swsusp_save
();
if
(
ret
==
0
)
_soft_restart
(
virt_to_
phys
(
cpu_resume
),
false
);
_soft_restart
(
virt_to_
idmap
(
cpu_resume
),
false
);
return
ret
;
}
...
...
@@ -87,7 +87,7 @@ static void notrace arch_restore_image(void *unused)
for
(
pbe
=
restore_pblist
;
pbe
;
pbe
=
pbe
->
next
)
copy_page
(
pbe
->
orig_address
,
pbe
->
address
);
_soft_restart
(
virt_to_
phys
(
cpu_resume
),
false
);
_soft_restart
(
virt_to_
idmap
(
cpu_resume
),
false
);
}
static
u64
resume_stack
[
PAGE_SIZE
/
2
/
sizeof
(
u64
)]
__nosavedata
;
...
...
arch/arm/kernel/hyp-stub.S
View file @
1b3bf847
...
...
@@ -17,6 +17,7 @@
*/
#include <linux/init.h>
#include <linux/irqchip/arm-gic-v3.h>
#include <linux/linkage.h>
#include <asm/assembler.h>
#include <asm/virt.h>
...
...
@@ -159,6 +160,29 @@ ARM_BE8(orr r7, r7, #(1 << 25)) @ HSCTLR.EE
bic
r7
,
#
1
@
Clear
ENABLE
mcr
p15
,
0
,
r7
,
c14
,
c3
,
1
@
CNTV_CTL
1
:
#endif
#ifdef CONFIG_ARM_GIC_V3
@
Check
whether
GICv3
system
registers
are
available
mrc
p15
,
0
,
r7
,
c0
,
c1
,
1
@
ID_PFR1
ubfx
r7
,
r7
,
#
28
,
#
4
cmp
r7
,
#
1
bne
2
f
@
Enable
system
register
accesses
mrc
p15
,
4
,
r7
,
c12
,
c9
,
5
@
ICC_HSRE
orr
r7
,
r7
,
#(
ICC_SRE_EL2_ENABLE
|
ICC_SRE_EL2_SRE
)
mcr
p15
,
4
,
r7
,
c12
,
c9
,
5
@
ICC_HSRE
isb
@
SRE
bit
could
be
forced
to
0
by
firmware
.
@
Check
whether
it
sticks
before
accessing
any
other
sysreg
mrc
p15
,
4
,
r7
,
c12
,
c9
,
5
@
ICC_HSRE
tst
r7
,
#
ICC_SRE_EL2_SRE
beq
2
f
mov
r7
,
#
0
mcr
p15
,
4
,
r7
,
c12
,
c11
,
0
@
ICH_HCR
2
:
#endif
bx
lr
@
The
boot
CPU
mode
is
left
in
r4
.
...
...
arch/arm/kernel/irq.c
View file @
1b3bf847
...
...
@@ -95,7 +95,7 @@ void __init init_IRQ(void)
outer_cache
.
write_sec
=
machine_desc
->
l2c_write_sec
;
ret
=
l2x0_of_init
(
machine_desc
->
l2c_aux_val
,
machine_desc
->
l2c_aux_mask
);
if
(
ret
)
if
(
ret
&&
ret
!=
-
ENODEV
)
pr_err
(
"L2C: failed to init: %d
\n
"
,
ret
);
}
...
...
arch/arm/kernel/machine_kexec.c
View file @
1b3bf847
...
...
@@ -143,10 +143,8 @@ void (*kexec_reinit)(void);
void
machine_kexec
(
struct
kimage
*
image
)
{
unsigned
long
page_list
;
unsigned
long
reboot_code_buffer_phys
;
unsigned
long
reboot_entry
=
(
unsigned
long
)
relocate_new_kernel
;
unsigned
long
reboot_entry_phys
;
unsigned
long
page_list
,
reboot_entry_phys
;
void
(
*
reboot_entry
)(
void
);
void
*
reboot_code_buffer
;
/*
...
...
@@ -159,9 +157,6 @@ void machine_kexec(struct kimage *image)
page_list
=
image
->
head
&
PAGE_MASK
;
/* we need both effective and real address here */
reboot_code_buffer_phys
=
page_to_pfn
(
image
->
control_code_page
)
<<
PAGE_SHIFT
;
reboot_code_buffer
=
page_address
(
image
->
control_code_page
);
/* Prepare parameters for reboot_code_buffer*/
...
...
@@ -174,10 +169,11 @@ void machine_kexec(struct kimage *image)
/* copy our kernel relocation code to the control code page */
reboot_entry
=
fncpy
(
reboot_code_buffer
,
reboot_entry
,
&
relocate_new_kernel
,
relocate_new_kernel_size
);
reboot_entry_phys
=
(
unsigned
long
)
reboot_entry
+
(
reboot_code_buffer_phys
-
(
unsigned
long
)
reboot_code_buffer
);
/* get the identity mapping physical address for the reboot code */
reboot_entry_phys
=
virt_to_idmap
(
reboot_entry
);
pr_info
(
"Bye!
\n
"
);
...
...
arch/arm/kernel/module.c
View file @
1b3bf847
...
...
@@ -34,7 +34,7 @@
* recompiling the whole kernel when CONFIG_XIP_KERNEL is turned on/off.
*/
#undef MODULES_VADDR
#define MODULES_VADDR (((unsigned long)_e
text
+ ~PMD_MASK) & PMD_MASK)
#define MODULES_VADDR (((unsigned long)_e
xiprom
+ ~PMD_MASK) & PMD_MASK)
#endif
#ifdef CONFIG_MMU
...
...
arch/arm/kernel/reboot.c
View file @
1b3bf847
...
...
@@ -50,7 +50,7 @@ static void __soft_restart(void *addr)
flush_cache_all
();
/* Switch to the identity mapping. */
phys_reset
=
(
phys_reset_t
)
(
unsigned
long
)
virt_to_idmap
(
cpu_reset
);
phys_reset
=
(
phys_reset_t
)
virt_to_idmap
(
cpu_reset
);
phys_reset
((
unsigned
long
)
addr
);
/* Should never get here. */
...
...
arch/arm/kernel/topology.c
View file @
1b3bf847
...
...
@@ -40,7 +40,7 @@
* to run the rebalance_domains for all idle cores and the cpu_capacity can be
* updated during this sequence.
*/
static
DEFINE_PER_CPU
(
unsigned
long
,
cpu_scale
);
static
DEFINE_PER_CPU
(
unsigned
long
,
cpu_scale
)
=
SCHED_CAPACITY_SCALE
;
unsigned
long
arch_scale_cpu_capacity
(
struct
sched_domain
*
sd
,
int
cpu
)
{
...
...
@@ -306,8 +306,6 @@ void __init init_cpu_topology(void)
cpu_topo
->
socket_id
=
-
1
;
cpumask_clear
(
&
cpu_topo
->
core_sibling
);
cpumask_clear
(
&
cpu_topo
->
thread_sibling
);
set_capacity_scale
(
cpu
,
SCHED_CAPACITY_SCALE
);
}
smp_wmb
();
...
...
arch/arm/kernel/vmlinux-xip.lds.S
0 → 100644
View file @
1b3bf847
/*
ld
script
to
make
ARM
Linux
kernel
*
taken
from
the
i386
version
by
Russell
King
*
Written
by
Martin
Mares
<
mj
@
atrey
.
karlin
.
mff
.
cuni
.
cz
>
*/
#include <asm-generic/vmlinux.lds.h>
#include <asm/cache.h>
#include <asm/thread_info.h>
#include <asm/memory.h>
#include <asm/page.h>
#define PROC_INFO \
.
=
ALIGN
(
4
)
; \
VMLINUX_SYMBOL
(
__proc_info_begin
)
=
.
; \
*(.
proc.info.init
)
\
VMLINUX_SYMBOL
(
__proc_info_end
)
=
.
;
#define IDMAP_TEXT \
ALIGN_FUNCTION
()
; \
VMLINUX_SYMBOL
(
__idmap_text_start
)
=
.
; \
*(.
idmap.text
)
\
VMLINUX_SYMBOL
(
__idmap_text_end
)
=
.
; \
.
=
ALIGN
(
PAGE_SIZE
)
; \
VMLINUX_SYMBOL
(
__hyp_idmap_text_start
)
=
.
; \
*(.
hyp.idmap.text
)
\
VMLINUX_SYMBOL
(
__hyp_idmap_text_end
)
=
.
;
#ifdef CONFIG_HOTPLUG_CPU
#define ARM_CPU_DISCARD(x)
#define ARM_CPU_KEEP(x) x
#else
#define ARM_CPU_DISCARD(x) x
#define ARM_CPU_KEEP(x)
#endif
#if (defined(CONFIG_SMP_ON_UP) && !defined(CONFIG_DEBUG_SPINLOCK)) || \
defined
(
CONFIG_GENERIC_BUG
)
#define ARM_EXIT_KEEP(x) x
#define ARM_EXIT_DISCARD(x)
#else
#define ARM_EXIT_KEEP(x)
#define ARM_EXIT_DISCARD(x) x
#endif
OUTPUT_ARCH
(
arm
)
ENTRY
(
stext
)
#ifndef __ARMEB__
jiffies
=
jiffies_64
;
#else
jiffies
=
jiffies_64
+
4
;
#endif
SECTIONS
{
/
*
*
XXX
:
The
linker
does
not
define
how
output
sections
are
*
assigned
to
input
sections
when
there
are
multiple
statements
*
matching
the
same
input
section
name
.
There
is
no
documented
*
order
of
matching
.
*
*
unwind
exit
sections
must
be
discarded
before
the
rest
of
the
*
unwind
sections
get
included
.
*/
/
DISCARD
/
:
{
*(.
ARM.exidx.exit.text
)
*(.
ARM.extab.exit.text
)
ARM_CPU_DISCARD
(*(.
ARM.exidx.cpuexit.text
))
ARM_CPU_DISCARD
(*(.
ARM.extab.cpuexit.text
))
ARM_EXIT_DISCARD
(
EXIT_TEXT
)
ARM_EXIT_DISCARD
(
EXIT_DATA
)
EXIT_CALL
#ifndef CONFIG_MMU
*(.
text.fixup
)
*(
__ex_table
)
#endif
#ifndef CONFIG_SMP_ON_UP
*(.
alt.smp.init
)
#endif
*(.
discard
)
*(.
discard
.*)
}
.
=
XIP_VIRT_ADDR
(
CONFIG_XIP_PHYS_ADDR
)
;
_xiprom
=
.
; /* XIP ROM area to be mapped */
.
head.text
:
{
_text
=
.
;
HEAD_TEXT
}
.
text
:
{
/*
Real
text
segment
*/
_stext
=
.
; /* Text and read-only data */
IDMAP_TEXT
__exception_text_start
=
.
;
*(.
exception.text
)
__exception_text_end
=
.
;
IRQENTRY_TEXT
TEXT_TEXT
SCHED_TEXT
LOCK_TEXT
KPROBES_TEXT
*(.
gnu.warning
)
*(.
glue_7
)
*(.
glue_7t
)
.
=
ALIGN
(
4
)
;
*(.
got
)
/*
Global
offset
table
*/
ARM_CPU_KEEP
(
PROC_INFO
)
}
RO_DATA
(
PAGE_SIZE
)
.
=
ALIGN
(
4
)
;
__ex_table
:
AT
(
ADDR
(
__ex_table
)
-
LOAD_OFFSET
)
{
__start___ex_table
=
.
;
#ifdef CONFIG_MMU
*(
__ex_table
)
#endif
__stop___ex_table
=
.
;
}
#ifdef CONFIG_ARM_UNWIND
/
*
*
Stack
unwinding
tables
*/
.
=
ALIGN
(
8
)
;
.
ARM.
unwind_idx
:
{
__start_unwind_idx
=
.
;
*(.
ARM.exidx
*)
__stop_unwind_idx
=
.
;
}
.
ARM.
unwind_tab
:
{
__start_unwind_tab
=
.
;
*(.
ARM.extab
*)
__stop_unwind_tab
=
.
;
}
#endif
NOTES
_etext
=
.
; /* End of text and rodata section */
/
*
*
The
vectors
and
stubs
are
relocatable
code
,
and
the
*
only
thing
that
matters
is
their
relative
offsets
*/
__vectors_start
=
.
;
.
vectors
0xffff0000
:
AT
(
__vectors_start
)
{
*(.
vectors
)
}
.
=
__vectors_start
+
SIZEOF
(
.
vectors
)
;
__vectors_end
=
.
;
__stubs_start
=
.
;
.
stubs
ADDR
(
.
vectors
)
+
0x1000
:
AT
(
__stubs_start
)
{
*(.
stubs
)
}
.
=
__stubs_start
+
SIZEOF
(
.
stubs
)
;
__stubs_end
=
.
;
PROVIDE
(
vector_fiq_offset
=
vector_fiq
-
ADDR
(
.
vectors
))
;
INIT_TEXT_SECTION
(8)
.
exit.text
:
{
ARM_EXIT_KEEP
(
EXIT_TEXT
)
}
.
init.proc.info
:
{
ARM_CPU_DISCARD
(
PROC_INFO
)
}
.
init.arch.info
:
{
__arch_info_begin
=
.
;
*(.
arch.info.init
)
__arch_info_end
=
.
;
}
.
init.tagtable
:
{
__tagtable_begin
=
.
;
*(.
taglist.init
)
__tagtable_end
=
.
;
}
#ifdef CONFIG_SMP_ON_UP
.
init.smpalt
:
{
__smpalt_begin
=
.
;
*(.
alt.smp.init
)
__smpalt_end
=
.
;
}
#endif
.
init.
pv_table
:
{
__pv_table_begin
=
.
;
*(.
pv_table
)
__pv_table_end
=
.
;
}
.
init.data
:
{
INIT_SETUP
(16)
INIT_CALLS
CON_INITCALL
SECURITY_INITCALL
INIT_RAM_FS
}
#ifdef CONFIG_SMP
PERCPU_SECTION
(
L1_CACHE_BYTES
)
#endif
_exiprom
=
.
; /* End of XIP ROM area */
__data_loc
=
ALIGN
(
4
)
; /* location in binary */
.
=
PAGE_OFFSET
+
TEXT_OFFSET
;
.
data
:
AT
(
__data_loc
)
{
_data
=
.
; /* address in memory */
_sdata
=
.
;
/
*
*
first
,
the
init
task
union
,
aligned
*
to
an
8192
byte
boundary
.
*/
INIT_TASK_DATA
(
THREAD_SIZE
)
.
=
ALIGN
(
PAGE_SIZE
)
;
__init_begin
=
.
;
INIT_DATA
ARM_EXIT_KEEP
(
EXIT_DATA
)
.
=
ALIGN
(
PAGE_SIZE
)
;
__init_end
=
.
;
NOSAVE_DATA
CACHELINE_ALIGNED_DATA
(
L1_CACHE_BYTES
)
READ_MOSTLY_DATA
(
L1_CACHE_BYTES
)
/
*
*
and
the
usual
data
section
*/
DATA_DATA
CONSTRUCTORS
_edata
=
.
;
}
_edata_loc
=
__data_loc
+
SIZEOF
(
.
data
)
;
#ifdef CONFIG_HAVE_TCM
/
*
*
We
align
everything
to
a
page
boundary
so
we
can
*
free
it
after
init
has
commenced
and
TCM
contents
have
*
been
copied
to
its
destination
.
*/
.
tcm_start
:
{
.
=
ALIGN
(
PAGE_SIZE
)
;
__tcm_start
=
.
;
__itcm_start
=
.
;
}
/
*
*
Link
these
to
the
ITCM
RAM
*
Put
VMA
to
the
TCM
address
and
LMA
to
the
common
RAM
*
and
we
'll upload the contents from RAM to TCM and free
*
the
used
RAM
after
that
.
*/
.
text_itcm
ITCM_OFFSET
:
AT
(
__itcm_start
)
{
__sitcm_text
=
.
;
*(.
tcm.text
)
*(.
tcm.rodata
)
.
=
ALIGN
(
4
)
;
__eitcm_text
=
.
;
}
/
*
*
Reset
the
dot
pointer
,
this
is
needed
to
create
the
*
relative
__dtcm_start
below
(
to
be
used
as
extern
in
code
)
.
*/
.
=
ADDR
(
.
tcm_start
)
+
SIZEOF
(
.
tcm_start
)
+
SIZEOF
(
.
text_itcm
)
;
.
dtcm_start
:
{
__dtcm_start
=
.
;
}
/
*
TODO
:
add
remainder
of
ITCM
as
well
,
that
can
be
used
for
data
!
*/
.
data_dtcm
DTCM_OFFSET
:
AT
(
__dtcm_start
)
{
.
=
ALIGN
(
4
)
;
__sdtcm_data
=
.
;
*(.
tcm.data
)
.
=
ALIGN
(
4
)
;
__edtcm_data
=
.
;
}
/
*
Reset
the
dot
pointer
or
the
linker
gets
confused
*/
.
=
ADDR
(
.
dtcm_start
)
+
SIZEOF
(
.
data_dtcm
)
;
/
*
End
marker
for
freeing
TCM
copy
in
linked
object
*/
.
tcm_end
:
AT
(
ADDR
(
.
dtcm_start
)
+
SIZEOF
(
.
data_dtcm
)){
.
=
ALIGN
(
PAGE_SIZE
)
;
__tcm_end
=
.
;
}
#endif
BSS_SECTION
(0,
0
,
0
)
_end
=
.
;
STABS_DEBUG
}
/*
*
These
must
never
be
empty
*
If
you
have
to
comment
these
two
assert
statements
out
,
your
*
binutils
is
too
old
(
for
other
reasons
as
well
)
*/
ASSERT
((
__proc_info_end
-
__proc_info_begin
),
"missing CPU support"
)
ASSERT
((
__arch_info_end
-
__arch_info_begin
),
"no machine record defined"
)
/*
*
The
HYP
init
code
can
't be more than a page long,
*
and
should
not
cross
a
page
boundary
.
*
The
above
comment
applies
as
well
.
*/
ASSERT
(
__hyp_idmap_text_end
-
(
__hyp_idmap_text_start
&
PAGE_MASK
)
<=
PAGE_SIZE
,
"
HYP
init
code
too
big
or
misaligned
")
arch/arm/kernel/vmlinux.lds.S
View file @
1b3bf847
...
...
@@ -3,14 +3,16 @@
*
Written
by
Martin
Mares
<
mj
@
atrey
.
karlin
.
mff
.
cuni
.
cz
>
*/
#ifdef CONFIG_XIP_KERNEL
#include "vmlinux-xip.lds.S"
#else
#include <asm-generic/vmlinux.lds.h>
#include <asm/cache.h>
#include <asm/thread_info.h>
#include <asm/memory.h>
#include <asm/page.h>
#ifdef CONFIG_ARM_KERNMEM_PERMS
#include <asm/pgtable.h>
#endif
#define PROC_INFO \
.
=
ALIGN
(
4
)
; \
...
...
@@ -84,17 +86,13 @@ SECTIONS
*(.
discard
.*)
}
#ifdef CONFIG_XIP_KERNEL
.
=
XIP_VIRT_ADDR
(
CONFIG_XIP_PHYS_ADDR
)
;
#else
.
=
PAGE_OFFSET
+
TEXT_OFFSET
;
#endif
.
head.text
:
{
_text
=
.
;
HEAD_TEXT
}
#ifdef CONFIG_
ARM_KERNMEM_PERMS
#ifdef CONFIG_
DEBUG_RODATA
.
=
ALIGN
(
1
<<
SECTION_SHIFT
)
;
#endif
...
...
@@ -117,7 +115,7 @@ SECTIONS
ARM_CPU_KEEP
(
PROC_INFO
)
}
#ifdef CONFIG_DEBUG_RODATA
#ifdef CONFIG_DEBUG_
ALIGN_
RODATA
.
=
ALIGN
(
1
<<
SECTION_SHIFT
)
;
#endif
RO_DATA
(
PAGE_SIZE
)
...
...
@@ -152,32 +150,33 @@ SECTIONS
_etext
=
.
; /* End of text and rodata section */
#ifndef CONFIG_XIP_KERNEL
# ifdef CONFIG_ARM_KERNMEM_PERMS
#ifdef CONFIG_DEBUG_RODATA
.
=
ALIGN
(
1
<<
SECTION_SHIFT
)
;
#
else
#else
.
=
ALIGN
(
PAGE_SIZE
)
;
# endif
__init_begin
=
.
;
#endif
__init_begin
=
.
;
/
*
*
The
vectors
and
stubs
are
relocatable
code
,
and
the
*
only
thing
that
matters
is
their
relative
offsets
*/
__vectors_start
=
.
;
.
vectors
0
:
AT
(
__vectors_start
)
{
.
vectors
0
xffff0000
:
AT
(
__vectors_start
)
{
*(.
vectors
)
}
.
=
__vectors_start
+
SIZEOF
(
.
vectors
)
;
__vectors_end
=
.
;
__stubs_start
=
.
;
.
stubs
0x1000
:
AT
(
__stubs_start
)
{
.
stubs
ADDR
(
.
vectors
)
+
0x1000
:
AT
(
__stubs_start
)
{
*(.
stubs
)
}
.
=
__stubs_start
+
SIZEOF
(
.
stubs
)
;
__stubs_end
=
.
;
PROVIDE
(
vector_fiq_offset
=
vector_fiq
-
ADDR
(
.
vectors
))
;
INIT_TEXT_SECTION
(8)
.
exit.text
:
{
ARM_EXIT_KEEP
(
EXIT_TEXT
)
...
...
@@ -208,37 +207,28 @@ SECTIONS
__pv_table_end
=
.
;
}
.
init.data
:
{
#ifndef CONFIG_XIP_KERNEL
INIT_DATA
#endif
INIT_SETUP
(16)
INIT_CALLS
CON_INITCALL
SECURITY_INITCALL
INIT_RAM_FS
}
#ifndef CONFIG_XIP_KERNEL
.
exit.data
:
{
ARM_EXIT_KEEP
(
EXIT_DATA
)
}
#endif
#ifdef CONFIG_SMP
PERCPU_SECTION
(
L1_CACHE_BYTES
)
#endif
#ifdef CONFIG_XIP_KERNEL
__data_loc
=
ALIGN
(
4
)
; /* location in binary */
.
=
PAGE_OFFSET
+
TEXT_OFFSET
;
#else
#ifdef CONFIG_ARM_KERNMEM_PERMS
#ifdef CONFIG_DEBUG_RODATA
.
=
ALIGN
(
1
<<
SECTION_SHIFT
)
;
#else
.
=
ALIGN
(
THREAD_SIZE
)
;
#endif
__init_end
=
.
;
__data_loc
=
.
;
#endif
.
data
:
AT
(
__data_loc
)
{
_data
=
.
; /* address in memory */
...
...
@@ -250,15 +240,6 @@ SECTIONS
*/
INIT_TASK_DATA
(
THREAD_SIZE
)
#ifdef CONFIG_XIP_KERNEL
.
=
ALIGN
(
PAGE_SIZE
)
;
__init_begin
=
.
;
INIT_DATA
ARM_EXIT_KEEP
(
EXIT_DATA
)
.
=
ALIGN
(
PAGE_SIZE
)
;
__init_end
=
.
;
#endif
NOSAVE_DATA
CACHELINE_ALIGNED_DATA
(
L1_CACHE_BYTES
)
READ_MOSTLY_DATA
(
L1_CACHE_BYTES
)
...
...
@@ -336,6 +317,15 @@ SECTIONS
STABS_DEBUG
}
#ifdef CONFIG_DEBUG_RODATA
/*
*
Without
CONFIG_DEBUG_ALIGN_RODATA
,
__start_rodata_section_aligned
will
*
be
the
first
section
-
aligned
location
after
__start_rodata
.
Otherwise
,
*
it
will
be
equal
to
__start_rodata
.
*/
__start_rodata_section_aligned
=
ALIGN
(
__start_rodata
,
1
<<
SECTION_SHIFT
)
;
#endif
/*
*
These
must
never
be
empty
*
If
you
have
to
comment
these
two
assert
statements
out
,
your
...
...
@@ -351,3 +341,5 @@ ASSERT((__arch_info_end - __arch_info_begin), "no machine record defined")
*/
ASSERT
(
__hyp_idmap_text_end
-
(
__hyp_idmap_text_start
&
PAGE_MASK
)
<=
PAGE_SIZE
,
"
HYP
init
code
too
big
or
misaligned
")
#endif /* CONFIG_XIP_KERNEL */
arch/arm/mach-davinci/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -30,7 +30,7 @@
u32
*
uart
;
/* PORT_16C550A, in polled non-fifo mode */
static
void
putc
(
char
c
)
static
inline
void
putc
(
char
c
)
{
if
(
!
uart
)
return
;
...
...
arch/arm/mach-footbridge/Kconfig
View file @
1b3bf847
...
...
@@ -68,7 +68,6 @@ config ARCH_NETWINDER
select ISA
select ISA_DMA
select PCI
select VIRT_TO_BUS
help
Say Y here if you intend to run this kernel on the Rebel.COM
NetWinder. Information about this machine can be found at:
...
...
arch/arm/mach-keystone/keystone.c
View file @
1b3bf847
...
...
@@ -63,7 +63,7 @@ static void __init keystone_init(void)
of_platform_populate
(
NULL
,
of_default_bus_match_table
,
NULL
,
NULL
);
}
static
phys_addr_t
keystone_virt_to_idmap
(
unsigned
long
x
)
static
unsigned
long
keystone_virt_to_idmap
(
unsigned
long
x
)
{
return
(
phys_addr_t
)(
x
)
-
CONFIG_PAGE_OFFSET
+
KEYSTONE_LOW_PHYS_START
;
}
...
...
arch/arm/mach-ks8695/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -17,7 +17,7 @@
#include <linux/io.h>
#include <mach/regs-uart.h>
static
void
putc
(
char
c
)
static
inline
void
putc
(
char
c
)
{
while
(
!
(
__raw_readl
((
void
__iomem
*
)
KS8695_UART_PA
+
KS8695_URLS
)
&
URLS_URTHRE
))
barrier
();
...
...
arch/arm/mach-netx/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -40,7 +40,7 @@
#define FR_BUSY (1<<3)
#define FR_TXFF (1<<5)
static
void
putc
(
char
c
)
static
inline
void
putc
(
char
c
)
{
unsigned
long
base
;
...
...
arch/arm/mach-omap1/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -45,7 +45,7 @@ static void set_omap_uart_info(unsigned char port)
*
uart_info
=
port
;
}
static
void
putc
(
int
c
)
static
inline
void
putc
(
int
c
)
{
if
(
!
uart_base
)
return
;
...
...
arch/arm/mach-rpc/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -76,7 +76,7 @@ int white;
/*
* This does not append a newline
*/
static
void
putc
(
int
c
)
static
inline
void
putc
(
int
c
)
{
extern
void
ll_write_char
(
char
*
,
char
c
,
char
white
);
int
x
,
y
;
...
...
arch/arm/mach-sa1100/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -19,7 +19,7 @@
#define UART(x) (*(volatile unsigned long *)(serial_port + (x)))
static
void
putc
(
int
c
)
static
inline
void
putc
(
int
c
)
{
unsigned
long
serial_port
;
...
...
arch/arm/mach-w90x900/include/mach/uncompress.h
View file @
1b3bf847
...
...
@@ -27,7 +27,7 @@
#define TX_DONE (UART_LSR_TEMT | UART_LSR_THRE)
static
volatile
u32
*
const
uart_base
=
(
u32
*
)
UART0_PA
;
static
void
putc
(
int
ch
)
static
inline
void
putc
(
int
ch
)
{
/* Check THRE and TEMT bits before we transmit the character.
*/
...
...
arch/arm/mm/Kconfig
View file @
1b3bf847
...
...
@@ -1037,24 +1037,26 @@ config ARCH_SUPPORTS_BIG_ENDIAN
This option specifies the architecture can support big endian
operation.
config ARM_KERNMEM_PERMS
bool "Restrict kernel memory permissions"
depends on MMU
help
If this is set, kernel memory other than kernel text (and rodata)
will be made non-executable. The tradeoff is that each region is
padded to section-size (1MiB) boundaries (because their permissions
are different and splitting the 1M pages into 4K ones causes TLB
performance problems), wasting memory.
config DEBUG_RODATA
bool "Make kernel text and rodata read-only"
depends on ARM_KERNMEM_PERMS
depends on MMU && !XIP_KERNEL
default y if CPU_V7
help
If this is set, kernel text and rodata memory will be made
read-only, and non-text kernel memory will be made non-executable.
The tradeoff is that each region is padded to section-size (1MiB)
boundaries (because their permissions are different and splitting
the 1M pages into 4K ones causes TLB performance problems), which
can waste memory.
config DEBUG_ALIGN_RODATA
bool "Make rodata strictly non-executable"
depends on DEBUG_RODATA
default y
help
If this is set,
kernel text and rodata will be made read-only
. This
is to help catch accidental or malicious attempts to change the
kernel's executable code. Additionally splits rodata from kernel
text so it can be made explicitly non-executable. This creates
another section-size padded region, so it can waste mor
e memory
space
while gaining the read-only protections
.
If this is set,
rodata will be made explicitly non-executable
. This
provides protection on the rare chance that attackers might find and
use ROP gadgets that exist in the rodata section. This adds an
additional section-aligned split of rodata from kernel text so it
can be made explicitly non-executable. This padding may wast
e memory
space
to gain the additional protection
.
arch/arm/mm/cache-tauros2.c
View file @
1b3bf847
...
...
@@ -22,6 +22,11 @@
#include <asm/cputype.h>
#include <asm/hardware/cache-tauros2.h>
/* CP15 PJ4 Control configuration register */
#define CCR_L2C_PREFETCH_DISABLE BIT(24)
#define CCR_L2C_ECC_ENABLE BIT(23)
#define CCR_L2C_WAY7_4_DISABLE BIT(21)
#define CCR_L2C_BURST8_ENABLE BIT(20)
/*
* When Tauros2 is used on a CPU that supports the v7 hierarchical
...
...
@@ -182,18 +187,18 @@ static void enable_extra_feature(unsigned int features)
u
=
read_extra_features
();
if
(
features
&
CACHE_TAUROS2_PREFETCH_ON
)
u
&=
~
0x01000000
;
u
&=
~
CCR_L2C_PREFETCH_DISABLE
;
else
u
|=
0x01000000
;
u
|=
CCR_L2C_PREFETCH_DISABLE
;
pr_info
(
"Tauros2: %s L2 prefetch.
\n
"
,
(
features
&
CACHE_TAUROS2_PREFETCH_ON
)
?
"Enabling"
:
"Disabling"
);
if
(
features
&
CACHE_TAUROS2_LINEFILL_BURST8
)
u
|=
0x00100000
;
u
|=
CCR_L2C_BURST8_ENABLE
;
else
u
&=
~
0x00100000
;
pr_info
(
"Tauros2: %s
line fill burt8
.
\n
"
,
u
&=
~
CCR_L2C_BURST8_ENABLE
;
pr_info
(
"Tauros2: %s
burst8 line fill
.
\n
"
,
(
features
&
CACHE_TAUROS2_LINEFILL_BURST8
)
?
"Enabling"
:
"Disabling"
);
...
...
@@ -287,16 +292,15 @@ void __init tauros2_init(unsigned int features)
node
=
of_find_matching_node
(
NULL
,
tauros2_ids
);
if
(
!
node
)
{
pr_info
(
"Not found marvell,tauros2-cache, disable it
\n
"
);
return
;
}
else
{
ret
=
of_property_read_u32
(
node
,
"marvell,tauros2-cache-features"
,
&
f
);
if
(
ret
)
{
pr_info
(
"Not found marvell,tauros-cache-features property, "
"disable extra features
\n
"
);
features
=
0
;
}
else
features
=
f
;
}
ret
=
of_property_read_u32
(
node
,
"marvell,tauros2-cache-features"
,
&
f
);
if
(
ret
)
{
pr_info
(
"Not found marvell,tauros-cache-features property, "
"disable extra features
\n
"
);
features
=
0
;
}
else
features
=
f
;
#endif
tauros2_internal_init
(
features
);
}
arch/arm/mm/dma-mapping.c
View file @
1b3bf847
...
...
@@ -42,6 +42,55 @@
#include "dma.h"
#include "mm.h"
struct
arm_dma_alloc_args
{
struct
device
*
dev
;
size_t
size
;
gfp_t
gfp
;
pgprot_t
prot
;
const
void
*
caller
;
bool
want_vaddr
;
};
struct
arm_dma_free_args
{
struct
device
*
dev
;
size_t
size
;
void
*
cpu_addr
;
struct
page
*
page
;
bool
want_vaddr
;
};
struct
arm_dma_allocator
{
void
*
(
*
alloc
)(
struct
arm_dma_alloc_args
*
args
,
struct
page
**
ret_page
);
void
(
*
free
)(
struct
arm_dma_free_args
*
args
);
};
struct
arm_dma_buffer
{
struct
list_head
list
;
void
*
virt
;
struct
arm_dma_allocator
*
allocator
;
};
static
LIST_HEAD
(
arm_dma_bufs
);
static
DEFINE_SPINLOCK
(
arm_dma_bufs_lock
);
static
struct
arm_dma_buffer
*
arm_dma_buffer_find
(
void
*
virt
)
{
struct
arm_dma_buffer
*
buf
,
*
found
=
NULL
;
unsigned
long
flags
;
spin_lock_irqsave
(
&
arm_dma_bufs_lock
,
flags
);
list_for_each_entry
(
buf
,
&
arm_dma_bufs
,
list
)
{
if
(
buf
->
virt
==
virt
)
{
list_del
(
&
buf
->
list
);
found
=
buf
;
break
;
}
}
spin_unlock_irqrestore
(
&
arm_dma_bufs_lock
,
flags
);
return
found
;
}
/*
* The DMA API is built upon the notion of "buffer ownership". A buffer
* is either exclusively owned by the CPU (and therefore may be accessed
...
...
@@ -592,7 +641,7 @@ static inline pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot)
#define __alloc_remap_buffer(dev, size, gfp, prot, ret, c, wv) NULL
#define __alloc_from_pool(size, ret_page) NULL
#define __alloc_from_contiguous(dev, size, prot, ret, c, wv) NULL
#define __free_from_pool(cpu_addr, size)
0
#define __free_from_pool(cpu_addr, size)
do { } while (0)
#define __free_from_contiguous(dev, page, cpu_addr, size, wv) do { } while (0)
#define __dma_free_remap(cpu_addr, size) do { } while (0)
...
...
@@ -610,7 +659,78 @@ static void *__alloc_simple_buffer(struct device *dev, size_t size, gfp_t gfp,
return
page_address
(
page
);
}
static
void
*
simple_allocator_alloc
(
struct
arm_dma_alloc_args
*
args
,
struct
page
**
ret_page
)
{
return
__alloc_simple_buffer
(
args
->
dev
,
args
->
size
,
args
->
gfp
,
ret_page
);
}
static
void
simple_allocator_free
(
struct
arm_dma_free_args
*
args
)
{
__dma_free_buffer
(
args
->
page
,
args
->
size
);
}
static
struct
arm_dma_allocator
simple_allocator
=
{
.
alloc
=
simple_allocator_alloc
,
.
free
=
simple_allocator_free
,
};
static
void
*
cma_allocator_alloc
(
struct
arm_dma_alloc_args
*
args
,
struct
page
**
ret_page
)
{
return
__alloc_from_contiguous
(
args
->
dev
,
args
->
size
,
args
->
prot
,
ret_page
,
args
->
caller
,
args
->
want_vaddr
);
}
static
void
cma_allocator_free
(
struct
arm_dma_free_args
*
args
)
{
__free_from_contiguous
(
args
->
dev
,
args
->
page
,
args
->
cpu_addr
,
args
->
size
,
args
->
want_vaddr
);
}
static
struct
arm_dma_allocator
cma_allocator
=
{
.
alloc
=
cma_allocator_alloc
,
.
free
=
cma_allocator_free
,
};
static
void
*
pool_allocator_alloc
(
struct
arm_dma_alloc_args
*
args
,
struct
page
**
ret_page
)
{
return
__alloc_from_pool
(
args
->
size
,
ret_page
);
}
static
void
pool_allocator_free
(
struct
arm_dma_free_args
*
args
)
{
__free_from_pool
(
args
->
cpu_addr
,
args
->
size
);
}
static
struct
arm_dma_allocator
pool_allocator
=
{
.
alloc
=
pool_allocator_alloc
,
.
free
=
pool_allocator_free
,
};
static
void
*
remap_allocator_alloc
(
struct
arm_dma_alloc_args
*
args
,
struct
page
**
ret_page
)
{
return
__alloc_remap_buffer
(
args
->
dev
,
args
->
size
,
args
->
gfp
,
args
->
prot
,
ret_page
,
args
->
caller
,
args
->
want_vaddr
);
}
static
void
remap_allocator_free
(
struct
arm_dma_free_args
*
args
)
{
if
(
args
->
want_vaddr
)
__dma_free_remap
(
args
->
cpu_addr
,
args
->
size
);
__dma_free_buffer
(
args
->
page
,
args
->
size
);
}
static
struct
arm_dma_allocator
remap_allocator
=
{
.
alloc
=
remap_allocator_alloc
,
.
free
=
remap_allocator_free
,
};
static
void
*
__dma_alloc
(
struct
device
*
dev
,
size_t
size
,
dma_addr_t
*
handle
,
gfp_t
gfp
,
pgprot_t
prot
,
bool
is_coherent
,
...
...
@@ -619,7 +739,16 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle,
u64
mask
=
get_coherent_dma_mask
(
dev
);
struct
page
*
page
=
NULL
;
void
*
addr
;
bool
want_vaddr
;
bool
allowblock
,
cma
;
struct
arm_dma_buffer
*
buf
;
struct
arm_dma_alloc_args
args
=
{
.
dev
=
dev
,
.
size
=
PAGE_ALIGN
(
size
),
.
gfp
=
gfp
,
.
prot
=
prot
,
.
caller
=
caller
,
.
want_vaddr
=
!
dma_get_attr
(
DMA_ATTR_NO_KERNEL_MAPPING
,
attrs
),
};
#ifdef CONFIG_DMA_API_DEBUG
u64
limit
=
(
mask
+
1
)
&
~
mask
;
...
...
@@ -633,6 +762,10 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle,
if
(
!
mask
)
return
NULL
;
buf
=
kzalloc
(
sizeof
(
*
buf
),
gfp
);
if
(
!
buf
)
return
NULL
;
if
(
mask
<
0xffffffffULL
)
gfp
|=
GFP_DMA
;
...
...
@@ -644,28 +777,37 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle,
* platform; see CONFIG_HUGETLBFS.
*/
gfp
&=
~
(
__GFP_COMP
);
args
.
gfp
=
gfp
;
*
handle
=
DMA_ERROR_CODE
;
size
=
PAGE_ALIGN
(
size
);
want_vaddr
=
!
dma_get_attr
(
DMA_ATTR_NO_KERNEL_MAPPING
,
attrs
);
if
(
nommu
())
addr
=
__alloc_simple_buffer
(
dev
,
size
,
gfp
,
&
page
);
else
if
(
dev_get_cma_area
(
dev
)
&&
(
gfp
&
__GFP_DIRECT_RECLAIM
))
addr
=
__alloc_from_contiguous
(
dev
,
size
,
prot
,
&
page
,
caller
,
want_vaddr
);
else
if
(
is_coherent
)
addr
=
__alloc_simple_buffer
(
dev
,
size
,
gfp
,
&
page
);
else
if
(
!
gfpflags_allow_blocking
(
gfp
))
addr
=
__alloc_from_pool
(
size
,
&
page
);
allowblock
=
gfpflags_allow_blocking
(
gfp
);
cma
=
allowblock
?
dev_get_cma_area
(
dev
)
:
false
;
if
(
cma
)
buf
->
allocator
=
&
cma_allocator
;
else
if
(
nommu
()
||
is_coherent
)
buf
->
allocator
=
&
simple_allocator
;
else
if
(
allowblock
)
buf
->
allocator
=
&
remap_allocator
;
else
addr
=
__alloc_remap_buffer
(
dev
,
size
,
gfp
,
prot
,
&
page
,
caller
,
want_vaddr
);
buf
->
allocator
=
&
pool_allocator
;
addr
=
buf
->
allocator
->
alloc
(
&
args
,
&
page
);
if
(
page
)
{
unsigned
long
flags
;
if
(
page
)
*
handle
=
pfn_to_dma
(
dev
,
page_to_pfn
(
page
));
buf
->
virt
=
args
.
want_vaddr
?
addr
:
page
;
spin_lock_irqsave
(
&
arm_dma_bufs_lock
,
flags
);
list_add
(
&
buf
->
list
,
&
arm_dma_bufs
);
spin_unlock_irqrestore
(
&
arm_dma_bufs_lock
,
flags
);
}
else
{
kfree
(
buf
);
}
return
want_vaddr
?
addr
:
page
;
return
args
.
want_vaddr
?
addr
:
page
;
}
/*
...
...
@@ -741,25 +883,21 @@ static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr,
bool
is_coherent
)
{
struct
page
*
page
=
pfn_to_page
(
dma_to_pfn
(
dev
,
handle
));
bool
want_vaddr
=
!
dma_get_attr
(
DMA_ATTR_NO_KERNEL_MAPPING
,
attrs
);
size
=
PAGE_ALIGN
(
size
);
if
(
nommu
())
{
__dma_free_buffer
(
page
,
size
);
}
else
if
(
!
is_coherent
&&
__free_from_pool
(
cpu_addr
,
size
))
{
struct
arm_dma_buffer
*
buf
;
struct
arm_dma_free_args
args
=
{
.
dev
=
dev
,
.
size
=
PAGE_ALIGN
(
size
),
.
cpu_addr
=
cpu_addr
,
.
page
=
page
,
.
want_vaddr
=
!
dma_get_attr
(
DMA_ATTR_NO_KERNEL_MAPPING
,
attrs
),
};
buf
=
arm_dma_buffer_find
(
cpu_addr
);
if
(
WARN
(
!
buf
,
"Freeing invalid buffer %p
\n
"
,
cpu_addr
))
return
;
}
else
if
(
!
dev_get_cma_area
(
dev
))
{
if
(
want_vaddr
&&
!
is_coherent
)
__dma_free_remap
(
cpu_addr
,
size
);
__dma_free_buffer
(
page
,
size
);
}
else
{
/*
* Non-atomic allocations cannot be freed with IRQs disabled
*/
WARN_ON
(
irqs_disabled
());
__free_from_contiguous
(
dev
,
page
,
cpu_addr
,
size
,
want_vaddr
);
}
buf
->
allocator
->
free
(
&
args
);
kfree
(
buf
);
}
void
arm_dma_free
(
struct
device
*
dev
,
size_t
size
,
void
*
cpu_addr
,
...
...
@@ -1122,6 +1260,9 @@ static inline void __free_iova(struct dma_iommu_mapping *mapping,
spin_unlock_irqrestore
(
&
mapping
->
lock
,
flags
);
}
/* We'll try 2M, 1M, 64K, and finally 4K; array must end with 0! */
static
const
int
iommu_order_array
[]
=
{
9
,
8
,
4
,
0
};
static
struct
page
**
__iommu_alloc_buffer
(
struct
device
*
dev
,
size_t
size
,
gfp_t
gfp
,
struct
dma_attrs
*
attrs
)
{
...
...
@@ -1129,6 +1270,7 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size,
int
count
=
size
>>
PAGE_SHIFT
;
int
array_size
=
count
*
sizeof
(
struct
page
*
);
int
i
=
0
;
int
order_idx
=
0
;
if
(
array_size
<=
PAGE_SIZE
)
pages
=
kzalloc
(
array_size
,
GFP_KERNEL
);
...
...
@@ -1154,6 +1296,10 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size,
return
pages
;
}
/* Go straight to 4K chunks if caller says it's OK. */
if
(
dma_get_attr
(
DMA_ATTR_ALLOC_SINGLE_PAGES
,
attrs
))
order_idx
=
ARRAY_SIZE
(
iommu_order_array
)
-
1
;
/*
* IOMMU can map any pages, so himem can also be used here
*/
...
...
@@ -1162,22 +1308,24 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size,
while
(
count
)
{
int
j
,
order
;
for
(
order
=
__fls
(
count
);
order
>
0
;
--
order
)
{
/*
* We do not want OOM killer to be invoked as long
* as we can fall back to single pages, so we force
* __GFP_NORETRY for orders higher than zero.
*/
pages
[
i
]
=
alloc_pages
(
gfp
|
__GFP_NORETRY
,
order
);
if
(
pages
[
i
])
break
;
order
=
iommu_order_array
[
order_idx
];
/* Drop down when we get small */
if
(
__fls
(
count
)
<
order
)
{
order_idx
++
;
continue
;
}
if
(
!
pages
[
i
])
{
/*
* Fall back to single page allocation.
* Might invoke OOM killer as last resort.
*/
if
(
order
)
{
/* See if it's easy to allocate a high-order chunk */
pages
[
i
]
=
alloc_pages
(
gfp
|
__GFP_NORETRY
,
order
);
/* Go down a notch at first sign of pressure */
if
(
!
pages
[
i
])
{
order_idx
++
;
continue
;
}
}
else
{
pages
[
i
]
=
alloc_pages
(
gfp
,
0
);
if
(
!
pages
[
i
])
goto
error
;
...
...
arch/arm/mm/idmap.c
View file @
1b3bf847
...
...
@@ -15,7 +15,7 @@
* page tables.
*/
pgd_t
*
idmap_pgd
;
phys_addr_t
(
*
arch_virt_to_idmap
)
(
unsigned
long
x
);
unsigned
long
(
*
arch_virt_to_idmap
)
(
unsigned
long
x
);
#ifdef CONFIG_ARM_LPAE
static
void
idmap_add_pmd
(
pud_t
*
pud
,
unsigned
long
addr
,
unsigned
long
end
,
...
...
arch/arm/mm/init.c
View file @
1b3bf847
...
...
@@ -572,8 +572,9 @@ void __init mem_init(void)
}
}
#ifdef CONFIG_
ARM_KERNMEM_PERMS
#ifdef CONFIG_
DEBUG_RODATA
struct
section_perm
{
const
char
*
name
;
unsigned
long
start
;
unsigned
long
end
;
pmdval_t
mask
;
...
...
@@ -581,9 +582,13 @@ struct section_perm {
pmdval_t
clear
;
};
/* First section-aligned location at or after __start_rodata. */
extern
char
__start_rodata_section_aligned
[];
static
struct
section_perm
nx_perms
[]
=
{
/* Make pages tables, etc before _stext RW (set NX). */
{
.
name
=
"pre-text NX"
,
.
start
=
PAGE_OFFSET
,
.
end
=
(
unsigned
long
)
_stext
,
.
mask
=
~
PMD_SECT_XN
,
...
...
@@ -591,26 +596,26 @@ static struct section_perm nx_perms[] = {
},
/* Make init RW (set NX). */
{
.
name
=
"init NX"
,
.
start
=
(
unsigned
long
)
__init_begin
,
.
end
=
(
unsigned
long
)
_sdata
,
.
mask
=
~
PMD_SECT_XN
,
.
prot
=
PMD_SECT_XN
,
},
#ifdef CONFIG_DEBUG_RODATA
/* Make rodata NX (set RO in ro_perms below). */
{
.
start
=
(
unsigned
long
)
__start_rodata
,
.
name
=
"rodata NX"
,
.
start
=
(
unsigned
long
)
__start_rodata_section_aligned
,
.
end
=
(
unsigned
long
)
__init_begin
,
.
mask
=
~
PMD_SECT_XN
,
.
prot
=
PMD_SECT_XN
,
},
#endif
};
#ifdef CONFIG_DEBUG_RODATA
static
struct
section_perm
ro_perms
[]
=
{
/* Make kernel code and rodata RX (set RO). */
{
.
name
=
"text/rodata RO"
,
.
start
=
(
unsigned
long
)
_stext
,
.
end
=
(
unsigned
long
)
__init_begin
,
#ifdef CONFIG_ARM_LPAE
...
...
@@ -623,7 +628,6 @@ static struct section_perm ro_perms[] = {
#endif
},
};
#endif
/*
* Updates section permissions only for the current mm (sections are
...
...
@@ -670,8 +674,8 @@ void set_section_perms(struct section_perm *perms, int n, bool set,
for
(
i
=
0
;
i
<
n
;
i
++
)
{
if
(
!
IS_ALIGNED
(
perms
[
i
].
start
,
SECTION_SIZE
)
||
!
IS_ALIGNED
(
perms
[
i
].
end
,
SECTION_SIZE
))
{
pr_err
(
"BUG: section %lx-%lx not aligned to %lx
\n
"
,
perms
[
i
].
start
,
perms
[
i
].
end
,
pr_err
(
"BUG:
%s
section %lx-%lx not aligned to %lx
\n
"
,
perms
[
i
].
name
,
perms
[
i
].
start
,
perms
[
i
].
end
,
SECTION_SIZE
);
continue
;
}
...
...
@@ -712,7 +716,6 @@ void fix_kernmem_perms(void)
stop_machine
(
__fix_kernmem_perms
,
NULL
,
NULL
);
}
#ifdef CONFIG_DEBUG_RODATA
int
__mark_rodata_ro
(
void
*
unused
)
{
update_sections_early
(
ro_perms
,
ARRAY_SIZE
(
ro_perms
));
...
...
@@ -735,11 +738,10 @@ void set_kernel_text_ro(void)
set_section_perms
(
ro_perms
,
ARRAY_SIZE
(
ro_perms
),
true
,
current
->
active_mm
);
}
#endif
/* CONFIG_DEBUG_RODATA */
#else
static
inline
void
fix_kernmem_perms
(
void
)
{
}
#endif
/* CONFIG_
ARM_KERNMEM_PERMS
*/
#endif
/* CONFIG_
DEBUG_RODATA
*/
void
free_tcmmem
(
void
)
{
...
...
arch/arm/mm/mmu.c
View file @
1b3bf847
...
...
@@ -1253,7 +1253,7 @@ static inline void prepare_page_table(void)
#ifdef CONFIG_XIP_KERNEL
/* The XIP kernel is mapped in the module area -- skip over it */
addr
=
((
unsigned
long
)
_e
text
+
PMD_SIZE
-
1
)
&
PMD_MASK
;
addr
=
((
unsigned
long
)
_e
xiprom
+
PMD_SIZE
-
1
)
&
PMD_MASK
;
#endif
for
(
;
addr
<
PAGE_OFFSET
;
addr
+=
PMD_SIZE
)
pmd_clear
(
pmd_off_k
(
addr
));
...
...
@@ -1335,7 +1335,7 @@ static void __init devicemaps_init(const struct machine_desc *mdesc)
#ifdef CONFIG_XIP_KERNEL
map
.
pfn
=
__phys_to_pfn
(
CONFIG_XIP_PHYS_ADDR
&
SECTION_MASK
);
map
.
virtual
=
MODULES_VADDR
;
map
.
length
=
((
unsigned
long
)
_e
text
-
map
.
virtual
+
~
SECTION_MASK
)
&
SECTION_MASK
;
map
.
length
=
((
unsigned
long
)
_e
xiprom
-
map
.
virtual
+
~
SECTION_MASK
)
&
SECTION_MASK
;
map
.
type
=
MT_ROM
;
create_mapping
(
&
map
);
#endif
...
...
@@ -1426,7 +1426,11 @@ static void __init kmap_init(void)
static
void
__init
map_lowmem
(
void
)
{
struct
memblock_region
*
reg
;
#ifdef CONFIG_XIP_KERNEL
phys_addr_t
kernel_x_start
=
round_down
(
__pa
(
_sdata
),
SECTION_SIZE
);
#else
phys_addr_t
kernel_x_start
=
round_down
(
__pa
(
_stext
),
SECTION_SIZE
);
#endif
phys_addr_t
kernel_x_end
=
round_up
(
__pa
(
__init_end
),
SECTION_SIZE
);
/* Map all the lowmem memory banks. */
...
...
arch/arm/mm/pageattr.c
View file @
1b3bf847
...
...
@@ -49,6 +49,9 @@ static int change_memory_common(unsigned long addr, int numpages,
WARN_ON_ONCE
(
1
);
}
if
(
!
numpages
)
return
0
;
if
(
start
<
MODULES_VADDR
||
start
>=
MODULES_END
)
return
-
EINVAL
;
...
...
arch/arm/mm/proc-v7.S
View file @
1b3bf847
...
...
@@ -487,7 +487,7 @@ __errata_finish:
.
align
2
__v7_setup_stack_ptr
:
.
word
__v7_setup_stack
-
.
.
word
PHYS_RELATIVE
(
__v7_setup_stack
,
.
)
ENDPROC
(
__v7_setup
)
.
bss
...
...
arch/arm/plat-orion/time.c
View file @
1b3bf847
...
...
@@ -18,6 +18,7 @@
#include <linux/irq.h>
#include <linux/sched_clock.h>
#include <plat/time.h>
#include <asm/delay.h>
/*
* MBus bridge block registers.
...
...
@@ -188,6 +189,15 @@ orion_time_set_base(void __iomem *_timer_base)
timer_base
=
_timer_base
;
}
static
unsigned
long
orion_delay_timer_read
(
void
)
{
return
~
readl
(
timer_base
+
TIMER0_VAL_OFF
);
}
static
struct
delay_timer
orion_delay_timer
=
{
.
read_current_timer
=
orion_delay_timer_read
,
};
void
__init
orion_time_init
(
void
__iomem
*
_bridge_base
,
u32
_bridge_timer1_clr_mask
,
unsigned
int
irq
,
unsigned
int
tclk
)
...
...
@@ -202,6 +212,9 @@ orion_time_init(void __iomem *_bridge_base, u32 _bridge_timer1_clr_mask,
ticks_per_jiffy
=
(
tclk
+
HZ
/
2
)
/
HZ
;
orion_delay_timer
.
freq
=
tclk
;
register_current_timer_delay
(
&
orion_delay_timer
);
/*
* Set scale and timer for sched_clock.
*/
...
...
arch/arm64/kernel/psci.c
View file @
1b3bf847
...
...
@@ -20,7 +20,6 @@
#include <linux/smp.h>
#include <linux/delay.h>
#include <linux/psci.h>
#include <linux/slab.h>
#include <uapi/linux/psci.h>
...
...
@@ -28,73 +27,6 @@
#include <asm/cpu_ops.h>
#include <asm/errno.h>
#include <asm/smp_plat.h>
#include <asm/suspend.h>
static
DEFINE_PER_CPU_READ_MOSTLY
(
u32
*
,
psci_power_state
);
static
int
__maybe_unused
cpu_psci_cpu_init_idle
(
unsigned
int
cpu
)
{
int
i
,
ret
,
count
=
0
;
u32
*
psci_states
;
struct
device_node
*
state_node
,
*
cpu_node
;
cpu_node
=
of_get_cpu_node
(
cpu
,
NULL
);
if
(
!
cpu_node
)
return
-
ENODEV
;
/*
* If the PSCI cpu_suspend function hook has not been initialized
* idle states must not be enabled, so bail out
*/
if
(
!
psci_ops
.
cpu_suspend
)
return
-
EOPNOTSUPP
;
/* Count idle states */
while
((
state_node
=
of_parse_phandle
(
cpu_node
,
"cpu-idle-states"
,
count
)))
{
count
++
;
of_node_put
(
state_node
);
}
if
(
!
count
)
return
-
ENODEV
;
psci_states
=
kcalloc
(
count
,
sizeof
(
*
psci_states
),
GFP_KERNEL
);
if
(
!
psci_states
)
return
-
ENOMEM
;
for
(
i
=
0
;
i
<
count
;
i
++
)
{
u32
state
;
state_node
=
of_parse_phandle
(
cpu_node
,
"cpu-idle-states"
,
i
);
ret
=
of_property_read_u32
(
state_node
,
"arm,psci-suspend-param"
,
&
state
);
if
(
ret
)
{
pr_warn
(
" * %s missing arm,psci-suspend-param property
\n
"
,
state_node
->
full_name
);
of_node_put
(
state_node
);
goto
free_mem
;
}
of_node_put
(
state_node
);
pr_debug
(
"psci-power-state %#x index %d
\n
"
,
state
,
i
);
if
(
!
psci_power_state_is_valid
(
state
))
{
pr_warn
(
"Invalid PSCI power state %#x
\n
"
,
state
);
ret
=
-
EINVAL
;
goto
free_mem
;
}
psci_states
[
i
]
=
state
;
}
/* Idle states parsed correctly, initialize per-cpu pointer */
per_cpu
(
psci_power_state
,
cpu
)
=
psci_states
;
return
0
;
free_mem:
kfree
(
psci_states
);
return
ret
;
}
static
int
__init
cpu_psci_cpu_init
(
unsigned
int
cpu
)
{
...
...
@@ -178,38 +110,11 @@ static int cpu_psci_cpu_kill(unsigned int cpu)
}
#endif
static
int
psci_suspend_finisher
(
unsigned
long
index
)
{
u32
*
state
=
__this_cpu_read
(
psci_power_state
);
return
psci_ops
.
cpu_suspend
(
state
[
index
-
1
],
virt_to_phys
(
cpu_resume
));
}
static
int
__maybe_unused
cpu_psci_cpu_suspend
(
unsigned
long
index
)
{
int
ret
;
u32
*
state
=
__this_cpu_read
(
psci_power_state
);
/*
* idle state index 0 corresponds to wfi, should never be called
* from the cpu_suspend operations
*/
if
(
WARN_ON_ONCE
(
!
index
))
return
-
EINVAL
;
if
(
!
psci_power_state_loses_context
(
state
[
index
-
1
]))
ret
=
psci_ops
.
cpu_suspend
(
state
[
index
-
1
],
0
);
else
ret
=
cpu_suspend
(
index
,
psci_suspend_finisher
);
return
ret
;
}
const
struct
cpu_operations
cpu_psci_ops
=
{
.
name
=
"psci"
,
#ifdef CONFIG_CPU_IDLE
.
cpu_init_idle
=
cpu_
psci_cpu_init_idle
,
.
cpu_suspend
=
cpu_psci_cpu_suspend
,
.
cpu_init_idle
=
psci_cpu_init_idle
,
.
cpu_suspend
=
psci_cpu_suspend_enter
,
#endif
.
cpu_init
=
cpu_psci_cpu_init
,
.
cpu_prepare
=
cpu_psci_cpu_prepare
,
...
...
drivers/base/dd.c
View file @
1b3bf847
...
...
@@ -560,6 +560,7 @@ static int __device_attach_driver(struct device_driver *drv, void *_data)
struct
device_attach_data
*
data
=
_data
;
struct
device
*
dev
=
data
->
dev
;
bool
async_allowed
;
int
ret
;
/*
* Check if device has already been claimed. This may
...
...
@@ -570,8 +571,17 @@ static int __device_attach_driver(struct device_driver *drv, void *_data)
if
(
dev
->
driver
)
return
-
EBUSY
;
if
(
!
driver_match_device
(
drv
,
dev
))
ret
=
driver_match_device
(
drv
,
dev
);
if
(
ret
==
0
)
{
/* no match */
return
0
;
}
else
if
(
ret
==
-
EPROBE_DEFER
)
{
dev_dbg
(
dev
,
"Device match requests probe deferral
\n
"
);
driver_deferred_probe_add
(
dev
);
}
else
if
(
ret
<
0
)
{
dev_dbg
(
dev
,
"Bus failed to match device: %d"
,
ret
);
return
ret
;
}
/* ret > 0 means positive match */
async_allowed
=
driver_allows_async_probing
(
drv
);
...
...
@@ -691,6 +701,7 @@ void device_initial_probe(struct device *dev)
static
int
__driver_attach
(
struct
device
*
dev
,
void
*
data
)
{
struct
device_driver
*
drv
=
data
;
int
ret
;
/*
* Lock device and try to bind to it. We drop the error
...
...
@@ -702,8 +713,17 @@ static int __driver_attach(struct device *dev, void *data)
* is an error.
*/
if
(
!
driver_match_device
(
drv
,
dev
))
ret
=
driver_match_device
(
drv
,
dev
);
if
(
ret
==
0
)
{
/* no match */
return
0
;
}
else
if
(
ret
==
-
EPROBE_DEFER
)
{
dev_dbg
(
dev
,
"Device match requests probe deferral
\n
"
);
driver_deferred_probe_add
(
dev
);
}
else
if
(
ret
<
0
)
{
dev_dbg
(
dev
,
"Bus failed to match device: %d"
,
ret
);
return
ret
;
}
/* ret > 0 means positive match */
if
(
dev
->
parent
)
/* Needed for USB */
device_lock
(
dev
->
parent
);
...
...
drivers/clk/clkdev.c
View file @
1b3bf847
...
...
@@ -353,11 +353,25 @@ void clkdev_drop(struct clk_lookup *cl)
}
EXPORT_SYMBOL
(
clkdev_drop
);
static
struct
clk_lookup
*
__clk_register_clkdev
(
struct
clk_hw
*
hw
,
const
char
*
con_id
,
const
char
*
dev_id
,
...)
{
struct
clk_lookup
*
cl
;
va_list
ap
;
va_start
(
ap
,
dev_id
);
cl
=
vclkdev_create
(
hw
,
con_id
,
dev_id
,
ap
);
va_end
(
ap
);
return
cl
;
}
/**
* clk_register_clkdev - register one clock lookup for a struct clk
* @clk: struct clk to associate with all clk_lookups
* @con_id: connection ID string on device
* @dev_id:
format
string describing device name
* @dev_id: string describing device name
*
* con_id or dev_id may be NULL as a wildcard, just as in the rest of
* clkdev.
...
...
@@ -368,17 +382,22 @@ EXPORT_SYMBOL(clkdev_drop);
* after clk_register().
*/
int
clk_register_clkdev
(
struct
clk
*
clk
,
const
char
*
con_id
,
const
char
*
dev_
fmt
,
...
)
const
char
*
dev_
id
)
{
struct
clk_lookup
*
cl
;
va_list
ap
;
if
(
IS_ERR
(
clk
))
return
PTR_ERR
(
clk
);
va_start
(
ap
,
dev_fmt
);
cl
=
vclkdev_create
(
__clk_get_hw
(
clk
),
con_id
,
dev_fmt
,
ap
);
va_end
(
ap
);
/*
* Since dev_id can be NULL, and NULL is handled specially, we must
* pass it as either a NULL format string, or with "%s".
*/
if
(
dev_id
)
cl
=
__clk_register_clkdev
(
__clk_get_hw
(
clk
),
con_id
,
"%s"
,
dev_id
);
else
cl
=
__clk_register_clkdev
(
__clk_get_hw
(
clk
),
con_id
,
NULL
);
return
cl
?
0
:
-
ENOMEM
;
}
...
...
drivers/firmware/psci.c
View file @
1b3bf847
...
...
@@ -14,6 +14,7 @@
#define pr_fmt(fmt) "psci: " fmt
#include <linux/arm-smccc.h>
#include <linux/cpuidle.h>
#include <linux/errno.h>
#include <linux/linkage.h>
#include <linux/of.h>
...
...
@@ -21,10 +22,12 @@
#include <linux/printk.h>
#include <linux/psci.h>
#include <linux/reboot.h>
#include <linux/slab.h>
#include <linux/suspend.h>
#include <uapi/linux/psci.h>
#include <asm/cpuidle.h>
#include <asm/cputype.h>
#include <asm/system_misc.h>
#include <asm/smp_plat.h>
...
...
@@ -244,6 +247,123 @@ static int __init psci_features(u32 psci_func_id)
psci_func_id
,
0
,
0
);
}
#ifdef CONFIG_CPU_IDLE
static
DEFINE_PER_CPU_READ_MOSTLY
(
u32
*
,
psci_power_state
);
static
int
psci_dt_cpu_init_idle
(
struct
device_node
*
cpu_node
,
int
cpu
)
{
int
i
,
ret
,
count
=
0
;
u32
*
psci_states
;
struct
device_node
*
state_node
;
/*
* If the PSCI cpu_suspend function hook has not been initialized
* idle states must not be enabled, so bail out
*/
if
(
!
psci_ops
.
cpu_suspend
)
return
-
EOPNOTSUPP
;
/* Count idle states */
while
((
state_node
=
of_parse_phandle
(
cpu_node
,
"cpu-idle-states"
,
count
)))
{
count
++
;
of_node_put
(
state_node
);
}
if
(
!
count
)
return
-
ENODEV
;
psci_states
=
kcalloc
(
count
,
sizeof
(
*
psci_states
),
GFP_KERNEL
);
if
(
!
psci_states
)
return
-
ENOMEM
;
for
(
i
=
0
;
i
<
count
;
i
++
)
{
u32
state
;
state_node
=
of_parse_phandle
(
cpu_node
,
"cpu-idle-states"
,
i
);
ret
=
of_property_read_u32
(
state_node
,
"arm,psci-suspend-param"
,
&
state
);
if
(
ret
)
{
pr_warn
(
" * %s missing arm,psci-suspend-param property
\n
"
,
state_node
->
full_name
);
of_node_put
(
state_node
);
goto
free_mem
;
}
of_node_put
(
state_node
);
pr_debug
(
"psci-power-state %#x index %d
\n
"
,
state
,
i
);
if
(
!
psci_power_state_is_valid
(
state
))
{
pr_warn
(
"Invalid PSCI power state %#x
\n
"
,
state
);
ret
=
-
EINVAL
;
goto
free_mem
;
}
psci_states
[
i
]
=
state
;
}
/* Idle states parsed correctly, initialize per-cpu pointer */
per_cpu
(
psci_power_state
,
cpu
)
=
psci_states
;
return
0
;
free_mem:
kfree
(
psci_states
);
return
ret
;
}
int
psci_cpu_init_idle
(
unsigned
int
cpu
)
{
struct
device_node
*
cpu_node
;
int
ret
;
cpu_node
=
of_get_cpu_node
(
cpu
,
NULL
);
if
(
!
cpu_node
)
return
-
ENODEV
;
ret
=
psci_dt_cpu_init_idle
(
cpu_node
,
cpu
);
of_node_put
(
cpu_node
);
return
ret
;
}
static
int
psci_suspend_finisher
(
unsigned
long
index
)
{
u32
*
state
=
__this_cpu_read
(
psci_power_state
);
return
psci_ops
.
cpu_suspend
(
state
[
index
-
1
],
virt_to_phys
(
cpu_resume
));
}
int
psci_cpu_suspend_enter
(
unsigned
long
index
)
{
int
ret
;
u32
*
state
=
__this_cpu_read
(
psci_power_state
);
/*
* idle state index 0 corresponds to wfi, should never be called
* from the cpu_suspend operations
*/
if
(
WARN_ON_ONCE
(
!
index
))
return
-
EINVAL
;
if
(
!
psci_power_state_loses_context
(
state
[
index
-
1
]))
ret
=
psci_ops
.
cpu_suspend
(
state
[
index
-
1
],
0
);
else
ret
=
cpu_suspend
(
index
,
psci_suspend_finisher
);
return
ret
;
}
/* ARM specific CPU idle operations */
#ifdef CONFIG_ARM
static
struct
cpuidle_ops
psci_cpuidle_ops
__initdata
=
{
.
suspend
=
psci_cpu_suspend_enter
,
.
init
=
psci_dt_cpu_init_idle
,
};
CPUIDLE_METHOD_OF_DECLARE
(
psci
,
"arm,psci"
,
&
psci_cpuidle_ops
);
#endif
#endif
static
int
psci_system_suspend
(
unsigned
long
unused
)
{
return
invoke_psci_fn
(
PSCI_FN_NATIVE
(
1
_0
,
SYSTEM_SUSPEND
),
...
...
drivers/media/v4l2-core/videobuf2-dma-contig.c
View file @
1b3bf847
...
...
@@ -23,13 +23,16 @@
struct
vb2_dc_conf
{
struct
device
*
dev
;
struct
dma_attrs
attrs
;
};
struct
vb2_dc_buf
{
struct
device
*
dev
;
void
*
vaddr
;
unsigned
long
size
;
void
*
cookie
;
dma_addr_t
dma_addr
;
struct
dma_attrs
attrs
;
enum
dma_data_direction
dma_dir
;
struct
sg_table
*
dma_sgt
;
struct
frame_vector
*
vec
;
...
...
@@ -131,7 +134,8 @@ static void vb2_dc_put(void *buf_priv)
sg_free_table
(
buf
->
sgt_base
);
kfree
(
buf
->
sgt_base
);
}
dma_free_coherent
(
buf
->
dev
,
buf
->
size
,
buf
->
vaddr
,
buf
->
dma_addr
);
dma_free_attrs
(
buf
->
dev
,
buf
->
size
,
buf
->
cookie
,
buf
->
dma_addr
,
&
buf
->
attrs
);
put_device
(
buf
->
dev
);
kfree
(
buf
);
}
...
...
@@ -147,14 +151,18 @@ static void *vb2_dc_alloc(void *alloc_ctx, unsigned long size,
if
(
!
buf
)
return
ERR_PTR
(
-
ENOMEM
);
buf
->
vaddr
=
dma_alloc_coherent
(
dev
,
size
,
&
buf
->
dma_addr
,
GFP_KERNEL
|
gfp_flags
);
if
(
!
buf
->
vaddr
)
{
buf
->
attrs
=
conf
->
attrs
;
buf
->
cookie
=
dma_alloc_attrs
(
dev
,
size
,
&
buf
->
dma_addr
,
GFP_KERNEL
|
gfp_flags
,
&
buf
->
attrs
);
if
(
!
buf
->
cookie
)
{
dev_err
(
dev
,
"dma_alloc_coherent of size %ld failed
\n
"
,
size
);
kfree
(
buf
);
return
ERR_PTR
(
-
ENOMEM
);
}
if
(
!
dma_get_attr
(
DMA_ATTR_NO_KERNEL_MAPPING
,
&
buf
->
attrs
))
buf
->
vaddr
=
buf
->
cookie
;
/* Prevent the device from being released while the buffer is used */
buf
->
dev
=
get_device
(
dev
);
buf
->
size
=
size
;
...
...
@@ -185,8 +193,8 @@ static int vb2_dc_mmap(void *buf_priv, struct vm_area_struct *vma)
*/
vma
->
vm_pgoff
=
0
;
ret
=
dma_mmap_
coherent
(
buf
->
dev
,
vma
,
buf
->
vaddr
,
buf
->
dma_addr
,
buf
->
size
);
ret
=
dma_mmap_
attrs
(
buf
->
dev
,
vma
,
buf
->
cookie
,
buf
->
dma_addr
,
buf
->
size
,
&
buf
->
attrs
);
if
(
ret
)
{
pr_err
(
"Remapping memory failed, error: %d
\n
"
,
ret
);
...
...
@@ -329,7 +337,7 @@ static void *vb2_dc_dmabuf_ops_kmap(struct dma_buf *dbuf, unsigned long pgnum)
{
struct
vb2_dc_buf
*
buf
=
dbuf
->
priv
;
return
buf
->
vaddr
+
pgnum
*
PAGE_SIZE
;
return
buf
->
vaddr
?
buf
->
vaddr
+
pgnum
*
PAGE_SIZE
:
NULL
;
}
static
void
*
vb2_dc_dmabuf_ops_vmap
(
struct
dma_buf
*
dbuf
)
...
...
@@ -368,8 +376,8 @@ static struct sg_table *vb2_dc_get_base_sgt(struct vb2_dc_buf *buf)
return
NULL
;
}
ret
=
dma_get_sgtable
(
buf
->
dev
,
sgt
,
buf
->
vaddr
,
buf
->
dma_addr
,
buf
->
size
);
ret
=
dma_get_sgtable
_attrs
(
buf
->
dev
,
sgt
,
buf
->
cookie
,
buf
->
dma_addr
,
buf
->
size
,
&
buf
->
attrs
);
if
(
ret
<
0
)
{
dev_err
(
buf
->
dev
,
"failed to get scatterlist from DMA API
\n
"
);
kfree
(
sgt
);
...
...
@@ -721,7 +729,8 @@ const struct vb2_mem_ops vb2_dma_contig_memops = {
};
EXPORT_SYMBOL_GPL
(
vb2_dma_contig_memops
);
void
*
vb2_dma_contig_init_ctx
(
struct
device
*
dev
)
void
*
vb2_dma_contig_init_ctx_attrs
(
struct
device
*
dev
,
struct
dma_attrs
*
attrs
)
{
struct
vb2_dc_conf
*
conf
;
...
...
@@ -730,10 +739,12 @@ void *vb2_dma_contig_init_ctx(struct device *dev)
return
ERR_PTR
(
-
ENOMEM
);
conf
->
dev
=
dev
;
if
(
attrs
)
conf
->
attrs
=
*
attrs
;
return
conf
;
}
EXPORT_SYMBOL_GPL
(
vb2_dma_contig_init_ctx
);
EXPORT_SYMBOL_GPL
(
vb2_dma_contig_init_ctx
_attrs
);
void
vb2_dma_contig_cleanup_ctx
(
void
*
alloc_ctx
)
{
...
...
drivers/nvdimm/bus.c
View file @
1b3bf847
...
...
@@ -62,7 +62,7 @@ static int nvdimm_bus_match(struct device *dev, struct device_driver *drv)
{
struct
nd_device_driver
*
nd_drv
=
to_nd_device_driver
(
drv
);
return
test_bit
(
to_nd_device_type
(
dev
),
&
nd_drv
->
type
);
return
!!
test_bit
(
to_nd_device_type
(
dev
),
&
nd_drv
->
type
);
}
static
struct
module
*
to_bus_provider
(
struct
device
*
dev
)
...
...
include/linux/clkdev.h
View file @
1b3bf847
...
...
@@ -44,8 +44,7 @@ struct clk_lookup *clkdev_create(struct clk *clk, const char *con_id,
void
clkdev_add_table
(
struct
clk_lookup
*
,
size_t
);
int
clk_add_alias
(
const
char
*
,
const
char
*
,
const
char
*
,
struct
device
*
);
int
clk_register_clkdev
(
struct
clk
*
,
const
char
*
,
const
char
*
,
...)
__printf
(
3
,
4
);
int
clk_register_clkdev
(
struct
clk
*
,
const
char
*
,
const
char
*
);
int
clk_register_clkdevs
(
struct
clk
*
,
struct
clk_lookup
*
,
size_t
);
#ifdef CONFIG_COMMON_CLK
...
...
include/linux/device.h
View file @
1b3bf847
...
...
@@ -70,8 +70,11 @@ extern void bus_remove_file(struct bus_type *, struct bus_attribute *);
* @dev_groups: Default attributes of the devices on the bus.
* @drv_groups: Default attributes of the device drivers on the bus.
* @match: Called, perhaps multiple times, whenever a new device or driver
* is added for this bus. It should return a nonzero value if the
* given device can be handled by the given driver.
* is added for this bus. It should return a positive value if the
* given device can be handled by the given driver and zero
* otherwise. It may also return error code if determining that
* the driver supports the device is not possible. In case of
* -EPROBE_DEFER it will queue the device for deferred probing.
* @uevent: Called when a device is added, removed, or a few other things
* that generate uevents to add the environment variables.
* @probe: Called when a new device or driver add to this bus, and callback
...
...
include/linux/dma-attrs.h
View file @
1b3bf847
...
...
@@ -18,6 +18,7 @@ enum dma_attr {
DMA_ATTR_NO_KERNEL_MAPPING
,
DMA_ATTR_SKIP_CPU_SYNC
,
DMA_ATTR_FORCE_CONTIGUOUS
,
DMA_ATTR_ALLOC_SINGLE_PAGES
,
DMA_ATTR_MAX
,
};
...
...
include/linux/psci.h
View file @
1b3bf847
...
...
@@ -24,6 +24,9 @@ bool psci_tos_resident_on(int cpu);
bool
psci_power_state_loses_context
(
u32
state
);
bool
psci_power_state_is_valid
(
u32
state
);
int
psci_cpu_init_idle
(
unsigned
int
cpu
);
int
psci_cpu_suspend_enter
(
unsigned
long
index
);
struct
psci_operations
{
int
(
*
cpu_suspend
)(
u32
state
,
unsigned
long
entry_point
);
int
(
*
cpu_off
)(
u32
state
);
...
...
include/media/videobuf2-dma-contig.h
View file @
1b3bf847
...
...
@@ -16,6 +16,8 @@
#include <media/videobuf2-v4l2.h>
#include <linux/dma-mapping.h>
struct
dma_attrs
;
static
inline
dma_addr_t
vb2_dma_contig_plane_dma_addr
(
struct
vb2_buffer
*
vb
,
unsigned
int
plane_no
)
{
...
...
@@ -24,7 +26,14 @@ vb2_dma_contig_plane_dma_addr(struct vb2_buffer *vb, unsigned int plane_no)
return
*
addr
;
}
void
*
vb2_dma_contig_init_ctx
(
struct
device
*
dev
);
void
*
vb2_dma_contig_init_ctx_attrs
(
struct
device
*
dev
,
struct
dma_attrs
*
attrs
);
static
inline
void
*
vb2_dma_contig_init_ctx
(
struct
device
*
dev
)
{
return
vb2_dma_contig_init_ctx_attrs
(
dev
,
NULL
);
}
void
vb2_dma_contig_cleanup_ctx
(
void
*
alloc_ctx
);
extern
const
struct
vb2_mem_ops
vb2_dma_contig_memops
;
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment