* tmp-917a9: ARM/vdso: Mark the vDSO code read-only after init x86/vdso: Mark the vDSO code read-only after init lkdtm: Verify that '__ro_after_init' works correctly arch: Introduce post-init read-only memory x86/mm: Always enable CONFIG_DEBUG_RODATA and remove the Kconfig option mm/init: Add 'rodata=off' boot cmdline parameter to disable read-only kernel mappings asm-generic: Consolidate mark_rodata_ro() Linux 4.4.6 ld-version: Fix awk regex compile failure target: Drop incorrect ABORT_TASK put for completed commands block: don't optimize for non-cloned bio in bio_get_last_bvec() MIPS: smp.c: Fix uninitialised temp_foreign_map MIPS: Fix build error when SMP is used without GIC ovl: fix getcwd() failure after unsuccessful rmdir ovl: copy new uid/gid into overlayfs runtime inode userfaultfd: don't block on the last VM updates at exit time powerpc/powernv: Fix OPAL_CONSOLE_FLUSH prototype and usages powerpc/powernv: Add a kmsg_dumper that flushes console output on panic powerpc: Fix dedotify for binutils >= 2.26 Revert "drm/radeon/pm: adjust display configuration after powerstate" drm/radeon: Fix error handling in radeon_flip_work_func. drm/amdgpu: Fix error handling in amdgpu_flip_work_func. Revert "drm/radeon: call hpd_irq_event on resume" x86/mm: Fix slow_virt_to_phys() for X86_PAE again gpu: ipu-v3: Do not bail out on missing optional port nodes mac80211: Fix Public Action frame RX in AP mode mac80211: check PN correctly for GCMP-encrypted fragmented MPDUs mac80211: minstrel_ht: fix a logic error in RTS/CTS handling mac80211: minstrel_ht: set default tx aggregation timeout to 0 mac80211: fix use of uninitialised values in RX aggregation mac80211: minstrel: Change expected throughput unit back to Kbps iwlwifi: mvm: inc pending frames counter also when txing non-sta can: gs_usb: fixed disconnect bug by removing erroneous use of kfree() cfg80211/wext: fix message ordering wext: fix message delay/ordering ovl: fix working on distributed fs as lower layer ovl: ignore lower entries when checking purity of non-directory entries ASoC: wm8958: Fix enum ctl accesses in a wrong type ASoC: wm8994: Fix enum ctl accesses in a wrong type ASoC: samsung: Use IRQ safe spin lock calls ASoC: dapm: Fix ctl value accesses in a wrong type ncpfs: fix a braino in OOM handling in ncp_fill_cache() jffs2: reduce the breakage on recovery from halfway failed rename() dmaengine: at_xdmac: fix residue computation tracing: Fix check for cpu online when event is disabled s390/dasd: fix diag 0x250 inline assembly s390/mm: four page table levels vs. fork KVM: MMU: fix reserved bit check for ept=0/CR0.WP=0/CR4.SMEP=1/EFER.NX=0 KVM: MMU: fix ept=0/pte.u=1/pte.w=0/CR0.WP=0/CR4.SMEP=1/EFER.NX=0 combo KVM: PPC: Book3S HV: Sanitize special-purpose register values on guest exit KVM: s390: correct fprs on SIGP (STOP AND) STORE STATUS KVM: VMX: disable PEBS before a guest entry kvm: cap halt polling at exactly halt_poll_ns PCI: Allow a NULL "parent" pointer in pci_bus_assign_domain_nr() ARM: OMAP2+: hwmod: Introduce ti,no-idle dt property ARM: dts: dra7: do not gate cpsw clock due to errata i877 ARM: mvebu: fix overlap of Crypto SRAM with PCIe memory window arm64: account for sparsemem section alignment when choosing vmemmap offset Linux 4.4.5 drm/amdgpu: fix topaz/tonga gmc assignment in 4.4 stable modules: fix longstanding /proc/kallsyms vs module insertion race. drm/i915: refine qemu south bridge detection drm/i915: more virtual south bridge detection block: get the 1st and last bvec via helpers block: check virt boundary in bio_will_gap() drm/amdgpu: Use drm_calloc_large for VM page_tables array thermal: cpu_cooling: fix out of bounds access in time_in_idle i2c: brcmstb: allocate correct amount of memory for regmap ubi: Fix out of bounds write in volume update code cxl: Fix PSL timebase synchronization detection MIPS: traps: Fix SIGFPE information leak from `do_ov' and `do_trap_or_bp' MIPS: scache: Fix scache init with invalid line size. USB: serial: option: add support for Quectel UC20 USB: serial: option: add support for Telit LE922 PID 0x1045 USB: qcserial: add Sierra Wireless EM74xx device ID USB: qcserial: add Dell Wireless 5809e Gobi 4G HSPA+ (rev3) USB: cp210x: Add ID for Parrot NMEA GPS Flight Recorder usb: chipidea: otg: change workqueue ci_otg as freezable ALSA: timer: Fix broken compat timer user status ioctl ALSA: hdspm: Fix zero-division ALSA: hdsp: Fix wrong boolean ctl value accesses ALSA: hdspm: Fix wrong boolean ctl value accesses ALSA: seq: oss: Don't drain at closing a client ALSA: pcm: Fix ioctls for X32 ABI ALSA: timer: Fix ioctls for X32 ABI ALSA: rawmidi: Fix ioctls X32 ABI ALSA: hda - Fix mic issues on Acer Aspire E1-472 ALSA: ctl: Fix ioctls for X32 ABI ALSA: usb-audio: Add a quirk for Plantronics DA45 adv7604: fix tx 5v detect regression dmaengine: pxa_dma: fix cyclic transfers Fix directory hardlinks from deleted directories jffs2: Fix page lock / f->sem deadlock Revert "jffs2: Fix lock acquisition order bug in jffs2_write_begin" Btrfs: fix loading of orphan roots leading to BUG_ON pata-rb532-cf: get rid of the irq_to_gpio() call tracing: Do not have 'comm' filter override event 'comm' field ata: ahci: don't mark HotPlugCapable Ports as external/removable PM / sleep / x86: Fix crash on graph trace through x86 suspend arm64: vmemmap: use virtual projection of linear region Adding Intel Lewisburg device IDs for SATA writeback: flush inode cgroup wb switches instead of pinning super_block block: bio: introduce helpers to get the 1st and last bvec libata: Align ata_device's id on a cacheline libata: fix HDIO_GET_32BIT ioctl drm/amdgpu: return from atombios_dp_get_dpcd only when error drm/amdgpu/gfx8: specify which engine to wait before vm flush drm/amdgpu: apply gfx_v8 fixes to gfx_v7 as well drm/amdgpu/pm: update current crtc info after setting the powerstate drm/radeon/pm: update current crtc info after setting the powerstate drm/ast: Fix incorrect register check for DRAM width target: Fix WRITE_SAME/DISCARD conversion to linux 512b sectors iommu/vt-d: Use BUS_NOTIFY_REMOVED_DEVICE in hotplug path iommu/amd: Fix boot warning when device 00:00.0 is not iommu covered iommu/amd: Apply workaround for ATS write permission check arm/arm64: KVM: Fix ioctl error handling KVM: x86: fix root cause for missed hardware breakpoints vfio: fix ioctl error handling Fix cifs_uniqueid_to_ino_t() function for s390x CIFS: Fix SMB2+ interim response processing for read requests cifs: fix out-of-bounds access in lease parsing fbcon: set a default value to blink interval kvm: x86: Update tsc multiplier on change. mips/kvm: fix ioctl error handling parisc: Fix ptrace syscall number and return value modification PCI: keystone: Fix MSI code that retrieves struct pcie_port pointer block: Initialize max_dev_sectors to 0 drm/amdgpu: mask out WC from BO on unsupported arches btrfs: async-thread: Fix a use-after-free error for trace btrfs: Fix no_space in write and rm loop Btrfs: fix deadlock running delayed iputs at transaction commit time drivers: sh: Restore legacy clock domain on SuperH platforms use ->d_seq to get coherency between ->d_inode and ->d_flags Linux 4.4.4 iwlwifi: mvm: don't allow sched scans without matches to be started iwlwifi: update and fix 7265 series PCI IDs iwlwifi: pcie: properly configure the debug buffer size for 8000 iwlwifi: dvm: fix WoWLAN security: let security modules use PTRACE_MODE_* with bitmasks IB/cma: Fix RDMA port validation for iWarp x86/irq: Plug vector cleanup race x86/irq: Call irq_force_move_complete with irq descriptor x86/irq: Remove outgoing CPU from vector cleanup mask x86/irq: Remove the cpumask allocation from send_cleanup_vector() x86/irq: Clear move_in_progress before sending cleanup IPI x86/irq: Remove offline cpus from vector cleanup x86/irq: Get rid of code duplication x86/irq: Copy vectormask instead of an AND operation x86/irq: Check vector allocation early x86/irq: Reorganize the search in assign_irq_vector x86/irq: Reorganize the return path in assign_irq_vector x86/irq: Do not use apic_chip_data.old_domain as temporary buffer x86/irq: Validate that irq descriptor is still active x86/irq: Fix a race in x86_vector_free_irqs() x86/irq: Call chip->irq_set_affinity in proper context x86/entry/compat: Add missing CLAC to entry_INT80_32 x86/mpx: Fix off-by-one comparison with nr_registers hpfs: don't truncate the file when delete fails do_last(): ELOOP failure exit should be done after leaving RCU mode should_follow_link(): validate ->d_seq after having decided to follow xen/pcifront: Fix mysterious crashes when NUMA locality information was extracted. xen/pciback: Save the number of MSI-X entries to be copied later. xen/pciback: Check PF instead of VF for PCI_COMMAND_MEMORY xen/scsiback: correct frontend counting xen/arm: correctly handle DMA mapping of compound pages ARM: at91/dt: fix typo in sama5d2 pinmux descriptions ARM: OMAP2+: Fix onenand initialization to avoid filesystem corruption do_last(): don't let a bogus return value from ->open() et.al. to confuse us kernel/resource.c: fix muxed resource handling in __request_region() sunrpc/cache: fix off-by-one in qword_get() tracing: Fix showing function event in available_events powerpc/eeh: Fix partial hotplug criterion KVM: x86: MMU: fix ubsan index-out-of-range warning KVM: x86: fix conversion of addresses to linear in 32-bit protected mode KVM: x86: fix missed hardware breakpoints KVM: arm/arm64: vgic: Ensure bitmaps are long enough KVM: async_pf: do not warn on page allocation failures of/irq: Fix msi-map calculation for nonzero rid-base NFSv4: Fix a dentry leak on alias use nfs: fix nfs_size_to_loff_t block: fix use-after-free in dio_bio_complete bio: return EINTR if copying to user space got interrupted i2c: i801: Adding Intel Lewisburg support for iTCO phy: core: fix wrong err handle for phy_power_on writeback: keep superblock pinned during cgroup writeback association switches cgroup: make sure a parent css isn't offlined before its children cpuset: make mm migration asynchronous PCI/AER: Flush workqueue on device remove to avoid use-after-free ARCv2: SMP: Emulate IPI to self using software triggered interrupt ARCv2: STAR 9000950267: Handle return from intr to Delay Slot #2 libata: fix sff host state machine locking while polling qla2xxx: Fix stale pointer access. spi: atmel: fix gpio chip-select in case of non-DT platform target: Fix race with SCF_SEND_DELAYED_TAS handling target: Fix remote-port TMR ABORT + se_cmd fabric stop target: Fix TAS handling for multi-session se_node_acls target: Fix LUN_RESET active TMR descriptor handling target: Fix LUN_RESET active I/O handling for ACK_KREF ALSA: hda - Fixing background noise on Dell Inspiron 3162 ALSA: hda - Apply clock gate workaround to Skylake, too Revert "workqueue: make sure delayed work run in local cpu" workqueue: handle NUMA_NO_NODE for unbound pool_workqueue lookup mac80211: Requeue work after scan complete for all VIF types. rfkill: fix rfkill_fop_read wait_event usage tick/nohz: Set the correct expiry when switching to nohz/lowres mode perf stat: Do not clean event's private stats cdc-acm:exclude Samsung phone 04e8:685d Revert "Staging: panel: usleep_range is preferred over udelay" Staging: speakup: Fix getting port information sd: Optimal I/O size is in bytes, not sectors libceph: don't spam dmesg with stray reply warnings libceph: use the right footer size when skipping a message libceph: don't bail early from try_read() when skipping a message libceph: fix ceph_msg_revoke() seccomp: always propagate NO_NEW_PRIVS on tsync cpufreq: Fix NULL reference crash while accessing policy->governor_data cpufreq: pxa2xx: fix pxa_cpufreq_change_voltage prototype hwmon: (ads1015) Handle negative conversion values correctly hwmon: (gpio-fan) Remove un-necessary speed_index lookup for thermal hook hwmon: (dell-smm) Blacklist Dell Studio XPS 8000 Thermal: do thermal zone update after a cooling device registered Thermal: handle thermal zone device properly during system sleep Thermal: initialize thermal zone device correctly IB/mlx5: Expose correct maximum number of CQE capacity IB/qib: Support creating qps with GFP_NOIO flag IB/qib: fix mcast detach when qp not attached IB/cm: Fix a recently introduced deadlock dmaengine: dw: disable BLOCK IRQs for non-cyclic xfer dmaengine: at_xdmac: fix resume for cyclic transfers dmaengine: dw: fix cyclic transfer callbacks dmaengine: dw: fix cyclic transfer setup nfit: fix multi-interface dimm handling, acpi6.1 compatibility ACPI / PCI / hotplug: unlock in error path in acpiphp_enable_slot() ACPI: Revert "ACPI / video: Add Dell Inspiron 5737 to the blacklist" ACPI / video: Add disable_backlight_sysfs_if quirk for the Toshiba Satellite R830 ACPI / video: Add disable_backlight_sysfs_if quirk for the Toshiba Portege R700 lib: sw842: select crc32 uapi: update install list after nvme.h rename ideapad-laptop: Add Lenovo Yoga 700 to no_hw_rfkill dmi list ideapad-laptop: Add Lenovo ideapad Y700-17ISK to no_hw_rfkill dmi list toshiba_acpi: Fix blank screen at boot if transflective backlight is supported make sure that freeing shmem fast symlinks is RCU-delayed drm/radeon/pm: adjust display configuration after powerstate drm/radeon: Don't hang in radeon_flip_work_func on disabled crtc. (v2) drm: Fix treatment of drm_vblank_offdelay in drm_vblank_on() (v2) drm: Fix drm_vblank_pre/post_modeset regression from Linux 4.4 drm: Prevent vblank counter bumps > 1 with active vblank clients. (v2) drm: No-Op redundant calls to drm_vblank_off() (v2) drm/radeon: use post-decrement in error handling drm/qxl: use kmalloc_array to alloc reloc_info in qxl_process_single_command drm/i915: fix error path in intel_setup_gmbus() drm/i915/dsi: don't pass arbitrary data to sideband drm/i915/dsi: defend gpio table against out of bounds access drm/i915/skl: Don't skip mst encoders in skl_ddi_pll_select() drm/i915: Don't reject primary plane windowing with color keying enabled on SKL+ drm/i915/dp: fall back to 18 bpp when sink capability is unknown drm/i915: Make sure DC writes are coherent on flush. drm/i915: Init power domains early in driver load drm/i915: intel_hpd_init(): Fix suspend/resume reprobing drm/i915: Restore inhibiting the load of the default context drm: fix missing reference counting decrease drm/radeon: hold reference to fences in radeon_sa_bo_new drm/radeon: mask out WC from BO on unsupported arches drm: add helper to check for wc memory support drm/radeon: fix DP audio support for APU with DCE4.1 display engine drm/radeon: Add a common function for DFS handling drm/radeon: cleaned up VCO output settings for DP audio drm/radeon: properly byte swap vce firmware setup drm/radeon: clean up fujitsu quirks drm/radeon: Fix "slow" audio over DP on DCE8+ drm/radeon: call hpd_irq_event on resume drm/radeon: Fix off-by-one errors in radeon_vm_bo_set_addr drm/dp/mst: deallocate payload on port destruction drm/dp/mst: Reverse order of MST enable and clearing VC payload table. drm/dp/mst: move GUID storage from mgr, port to only mst branch drm/dp/mst: Calculate MST PBN with 31.32 fixed point drm: Add drm_fixp_from_fraction and drm_fixp2int_ceil drm/dp/mst: fix in RAD element access drm/dp/mst: fix in MSTB RAD initialization drm/dp/mst: always send reply for UP request drm/dp/mst: process broadcast messages correctly drm/nouveau: platform: Fix deferred probe drm/nouveau/disp/dp: ensure sink is powered up before attempting link training drm/nouveau/display: Enable vblank irqs after display engine is on again. drm/nouveau/kms: take mode_config mutex in connector hotplug path drm/amdgpu/pm: adjust display configuration after powerstate drm/amdgpu: Don't hang in amdgpu_flip_work_func on disabled crtc. drm/amdgpu: use post-decrement in error handling drm/amdgpu: fix issue with overlapping userptrs drm/amdgpu: hold reference to fences in amdgpu_sa_bo_new (v2) drm/amdgpu: remove unnecessary forward declaration drm/amdgpu: fix s4 resume drm/amdgpu: remove exp hardware support from iceland drm/amdgpu: don't load MEC2 on topaz drm/amdgpu: drop topaz support from gmc8 module drm/amdgpu: pull topaz gmc bits into gmc_v7 drm/amdgpu: The VI specific EXE bit should only apply to GMC v8.0 above drm/amdgpu: iceland use CI based MC IP drm/amdgpu: move gmc7 support out of CIK dependency drm/amdgpu: no need to load MC firmware on fiji drm/amdgpu: fix amdgpu_bo_pin_restricted VRAM placing v2 drm/amdgpu: fix tonga smu resume drm/amdgpu: fix lost sync_to if scheduler is enabled. drm/amdgpu: call hpd_irq_event on resume drm/amdgpu: Fix off-by-one errors in amdgpu_vm_bo_map drm/vmwgfx: respect 'nomodeset' drm/vmwgfx: Fix a width / pitch mismatch on framebuffer updates drm/vmwgfx: Fix an incorrect lock check virtio_pci: fix use after free on release virtio_balloon: fix race between migration and ballooning virtio_balloon: fix race by fill and leak regulator: mt6311: MT6311_REGULATOR needs to select REGMAP_I2C regulator: axp20x: Fix GPIO LDO enable value for AXP22x clk: exynos: use irqsave version of spin_lock to avoid deadlock with irqs cxl: use correct operator when writing pcie config space values sparc64: fix incorrect sign extension in sys_sparc64_personality EDAC, mc_sysfs: Fix freeing bus' name EDAC: Robustify workqueues destruction MIPS: Fix buffer overflow in syscall_get_arguments() MIPS: Fix some missing CONFIG_CPU_MIPSR6 #ifdefs MIPS: hpet: Choose a safe value for the ETIME check MIPS: Loongson-3: Fix SMP_ASK_C0COUNT IPI handler Revert "MIPS: Fix PAGE_MASK definition" cputime: Prevent 32bit overflow in time[val|spec]_to_cputime() time: Avoid signed overflow in timekeeping_get_ns() Bluetooth: 6lowpan: Fix handling of uncompressed IPv6 packets Bluetooth: 6lowpan: Fix kernel NULL pointer dereferences Bluetooth: Fix incorrect removing of IRKs Bluetooth: Add support of Toshiba Broadcom based devices Bluetooth: Use continuous scanning when creating LE connections Drivers: hv: vmbus: Fix a Host signaling bug tools: hv: vss: fix the write()'s argument: error -> vss_msg mmc: sdhci: Allow override of get_cd() called from sdhci_request() mmc: sdhci: Allow override of mmc host operations mmc: sdhci-pci: Fix card detect race for Intel BXT/APL mmc: pxamci: fix again read-only gpio detection polarity mmc: sdhci-acpi: Fix card detect race for Intel BXT/APL mmc: mmci: fix an ages old detection error mmc: core: Enable tuning according to the actual timing mmc: sdhci: Fix sdhci_runtime_pm_bus_on/off() mmc: mmc: Fix incorrect use of driver strength switching HS200 and HS400 mmc: sdio: Fix invalid vdd in voltage switch power cycle mmc: sdhci: Fix DMA descriptor with zero data length mmc: sdhci-pci: Do not default to 33 Ohm driver strength for Intel SPT mmc: usdhi6rol0: handle NULL data in timeout clockevents/tcb_clksrc: Prevent disabling an already disabled clock posix-clock: Fix return code on the poll method's error path irqchip/gic-v3-its: Fix double ICC_EOIR write for LPI in EOImode==1 irqchip/atmel-aic: Fix wrong bit operation for IRQ priority irqchip/mxs: Add missing set_handle_irq() irqchip/omap-intc: Add support for spurious irq handling coresight: checking for NULL string in coresight_name_match() dm: fix dm_rq_target_io leak on faults with .request_fn DM w/ blk-mq paths dm snapshot: fix hung bios when copy error occurs dm space map metadata: remove unused variable in brb_pop() tda1004x: only update the frontend properties if locked vb2: fix a regression in poll() behavior for output,streams gspca: ov534/topro: prevent a division by 0 si2157: return -EINVAL if firmware blob is too big media: dvb-core: Don't force CAN_INVERSION_AUTO in oneshot mode rc: sunxi-cir: Initialize the spinlock properly namei: ->d_inode of a pinned dentry is stable only for positives mei: validate request value in client notify request ioctl mei: fix fasync return value on error rtlwifi: rtl8723be: Fix module parameter initialization rtlwifi: rtl8188ee: Fix module parameter initialization rtlwifi: rtl8192se: Fix module parameter initialization rtlwifi: rtl8723ae: Fix initialization of module parameters rtlwifi: rtl8192de: Fix incorrect module parameter descriptions rtlwifi: rtl8192ce: Fix handling of module parameters rtlwifi: rtl8192cu: Add missing parameter setup rtlwifi: rtl_pci: Fix kernel panic locks: fix unlock when fcntl_setlk races with a close um: link with -lpthread uml: fix hostfs mknod() uml: flush stdout before forking s390/fpu: signals vs. floating point control register s390/compat: correct restore of high gprs on signal return s390/dasd: fix performance drop s390/dasd: fix refcount for PAV reassignment s390/dasd: prevent incorrect length error under z/VM after PAV changes s390: fix normalization bug in exception table sorting btrfs: initialize the seq counter in struct btrfs_device Btrfs: Initialize btrfs_root->highest_objectid when loading tree root and subvolume roots Btrfs: fix transaction handle leak on failure to create hard link Btrfs: fix number of transaction units required to create symlink Btrfs: send, don't BUG_ON() when an empty symlink is found btrfs: statfs: report zero available if metadata are exhausted Btrfs: igrab inode in writepage Btrfs: add missing brelse when superblock checksum fails KVM: s390: fix memory overwrites when vx is disabled s390/kvm: remove dependency on struct save_area definition clocksource/drivers/vt8500: Increase the minimum delta genirq: Validate action before dereferencing it in handle_irq_event_percpu() mm: numa: quickly fail allocations for NUMA balancing on full nodes mm: thp: fix SMP race condition between THP page fault and MADV_DONTNEED ocfs2: unlock inode if deleting inode from orphan fails drm/i915: shut up gen8+ SDE irq dmesg noise iw_cxgb3: Fix incorrectly returning error on success spi: omap2-mcspi: Prevent duplicate gpio_request drivers: android: correct the size of struct binder_uintptr_t for BC_DEAD_BINDER_DONE USB: option: add "4G LTE usb-modem U901" USB: option: add support for SIM7100E USB: cp210x: add IDs for GE B650V3 and B850V3 boards usb: dwc3: Fix assignment of EP transfer resources can: ems_usb: Fix possible tx overflow dm thin: fix race condition when destroying thin pool workqueue bcache: Change refill_dirty() to always scan entire disk if necessary bcache: prevent crash on changing writeback_running bcache: allows use of register in udev to avoid "device_busy" error. bcache: unregister reboot notifier if bcache fails to unregister device bcache: fix a leak in bch_cached_dev_run() bcache: clear BCACHE_DEV_UNLINK_DONE flag when attaching a backing device bcache: Add a cond_resched() call to gc bcache: fix a livelock when we cause a huge number of cache misses lib/ucs2_string: Correct ucs2 -> utf8 conversion efi: Add pstore variables to the deletion whitelist efi: Make efivarfs entries immutable by default efi: Make our variable validation list include the guid efi: Do variable name validation tests in utf8 efi: Use ucs2_as_utf8 in efivarfs instead of open coding a bad version lib/ucs2_string: Add ucs2 -> utf8 helper functions ARM: 8457/1: psci-smp is built only for SMP drm/gma500: Use correct unref in the gem bo create function devm_memremap: Fix error value when memremap failed KVM: s390: fix guest fprs memory leak arm64: errata: Add -mpc-relative-literal-loads to build flags ARM: debug-ll: fix BCM63xx entry for multiplatform ext4: fix bh->b_state corruption sctp: Fix port hash table size computation unix_diag: fix incorrect sign extension in unix_lookup_by_ino tipc: unlock in error path rtnl: RTM_GETNETCONF: fix wrong return value IFF_NO_QUEUE: Fix for drivers not calling ether_setup() tcp/dccp: fix another race at listener dismantle route: check and remove route cache when we get route net_sched fix: reclassification needs to consider ether protocol changes pppoe: fix reference counting in PPPoE proxy l2tp: Fix error creating L2TP tunnels net/mlx4_en: Avoid changing dev->features directly in run-time net/mlx4_en: Choose time-stamping shift value according to HW frequency net/mlx4_en: Count HW buffer overrun only once qmi_wwan: add "4G LTE usb-modem U901" tcp: md5: release request socket instead of listener tipc: fix premature addition of node to lookup table af_unix: Guard against other == sk in unix_dgram_sendmsg af_unix: Don't set err in unix_stream_read_generic unless there was an error ipv4: fix memory leaks in ip_cmsg_send() callers bonding: Fix ARP monitor validation bpf: fix branch offset adjustment on backjumps after patching ctx expansion flow_dissector: Fix unaligned access in __skb_flow_dissector when used by eth_get_headlen net: Copy inner L3 and L4 headers as unaligned on GRE TEB sctp: translate network order to host order when users get a hmacid enic: increment devcmd2 result ring in case of timeout tg3: Fix for tg3 transmit queue 0 timed out when too many gso_segs net:Add sysctl_max_skb_frags tcp: do not drop syn_recv on all icmp reports unix: correctly track in-flight fds in sending process user_struct ipv6: fix a lockdep splat ipv6: addrconf: Fix recursive spin lock call ipv6/udp: use sticky pktinfo egress ifindex on connect() ipv6: enforce flowi6_oif usage in ip6_dst_lookup_tail() tcp: beware of alignments in tcp_get_info() switchdev: Require RTNL mutex to be held when sending FDB notifications inet: frag: Always orphan skbs inside ip_defrag() tipc: fix connection abort during subscription cancel net: dsa: fix mv88e6xxx switches sctp: allow setting SCTP_SACK_IMMEDIATELY by the application pptp: fix illegal memory access caused by multiple bind()s af_unix: fix struct pid memory leak tcp: fix NULL deref in tcp_v4_send_ack() lwt: fix rx checksum setting for lwt devices tunneling over ipv6 tunnels: Allow IPv6 UDP checksums to be correctly controlled. net: dp83640: Fix tx timestamp overflow handling. gro: Make GRO aware of lightweight tunnels. af_iucv: Validate socket address length in iucv_sock_bind() Conflicts: arch/arm64/Makefile arch/arm64/include/asm/cacheflush.h drivers/mmc/host/sdhci.c drivers/usb/dwc3/ep0.c drivers/usb/dwc3/gadget.c kernel/module.c sound/core/pcm_compat.c CRs-Fixed: 1010239 Signed-off-by: Runmin Wang <runminw@codeaurora.org> Change-Id: I41a28636fc9ad91f9d979b191784609476294cdf
859 lines
27 KiB
C
859 lines
27 KiB
C
/*
|
|
* Functions related to setting various queue properties from drivers
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/bio.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/bootmem.h> /* for max_pfn/max_low_pfn */
|
|
#include <linux/gcd.h>
|
|
#include <linux/lcm.h>
|
|
#include <linux/jiffies.h>
|
|
#include <linux/gfp.h>
|
|
|
|
#include "blk.h"
|
|
|
|
unsigned long blk_max_low_pfn;
|
|
EXPORT_SYMBOL(blk_max_low_pfn);
|
|
|
|
unsigned long blk_max_pfn;
|
|
|
|
/**
|
|
* blk_queue_prep_rq - set a prepare_request function for queue
|
|
* @q: queue
|
|
* @pfn: prepare_request function
|
|
*
|
|
* It's possible for a queue to register a prepare_request callback which
|
|
* is invoked before the request is handed to the request_fn. The goal of
|
|
* the function is to prepare a request for I/O, it can be used to build a
|
|
* cdb from the request data for instance.
|
|
*
|
|
*/
|
|
void blk_queue_prep_rq(struct request_queue *q, prep_rq_fn *pfn)
|
|
{
|
|
q->prep_rq_fn = pfn;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_prep_rq);
|
|
|
|
/**
|
|
* blk_queue_unprep_rq - set an unprepare_request function for queue
|
|
* @q: queue
|
|
* @ufn: unprepare_request function
|
|
*
|
|
* It's possible for a queue to register an unprepare_request callback
|
|
* which is invoked before the request is finally completed. The goal
|
|
* of the function is to deallocate any data that was allocated in the
|
|
* prepare_request callback.
|
|
*
|
|
*/
|
|
void blk_queue_unprep_rq(struct request_queue *q, unprep_rq_fn *ufn)
|
|
{
|
|
q->unprep_rq_fn = ufn;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_unprep_rq);
|
|
|
|
void blk_queue_softirq_done(struct request_queue *q, softirq_done_fn *fn)
|
|
{
|
|
q->softirq_done_fn = fn;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_softirq_done);
|
|
|
|
void blk_queue_rq_timeout(struct request_queue *q, unsigned int timeout)
|
|
{
|
|
q->rq_timeout = timeout;
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_rq_timeout);
|
|
|
|
void blk_queue_rq_timed_out(struct request_queue *q, rq_timed_out_fn *fn)
|
|
{
|
|
q->rq_timed_out_fn = fn;
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_rq_timed_out);
|
|
|
|
void blk_queue_lld_busy(struct request_queue *q, lld_busy_fn *fn)
|
|
{
|
|
q->lld_busy_fn = fn;
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_lld_busy);
|
|
|
|
/**
|
|
* blk_set_default_limits - reset limits to default values
|
|
* @lim: the queue_limits structure to reset
|
|
*
|
|
* Description:
|
|
* Returns a queue_limit struct to its default state.
|
|
*/
|
|
void blk_set_default_limits(struct queue_limits *lim)
|
|
{
|
|
lim->max_segments = BLK_MAX_SEGMENTS;
|
|
lim->max_integrity_segments = 0;
|
|
lim->seg_boundary_mask = BLK_SEG_BOUNDARY_MASK;
|
|
lim->virt_boundary_mask = 0;
|
|
lim->max_segment_size = BLK_MAX_SEGMENT_SIZE;
|
|
lim->max_sectors = lim->max_hw_sectors = BLK_SAFE_MAX_SECTORS;
|
|
lim->max_dev_sectors = 0;
|
|
lim->chunk_sectors = 0;
|
|
lim->max_write_same_sectors = 0;
|
|
lim->max_discard_sectors = 0;
|
|
lim->max_hw_discard_sectors = 0;
|
|
lim->discard_granularity = 0;
|
|
lim->discard_alignment = 0;
|
|
lim->discard_misaligned = 0;
|
|
lim->discard_zeroes_data = 0;
|
|
lim->logical_block_size = lim->physical_block_size = lim->io_min = 512;
|
|
lim->bounce_pfn = (unsigned long)(BLK_BOUNCE_ANY >> PAGE_SHIFT);
|
|
lim->alignment_offset = 0;
|
|
lim->io_opt = 0;
|
|
lim->misaligned = 0;
|
|
lim->cluster = 1;
|
|
}
|
|
EXPORT_SYMBOL(blk_set_default_limits);
|
|
|
|
/**
|
|
* blk_set_stacking_limits - set default limits for stacking devices
|
|
* @lim: the queue_limits structure to reset
|
|
*
|
|
* Description:
|
|
* Returns a queue_limit struct to its default state. Should be used
|
|
* by stacking drivers like DM that have no internal limits.
|
|
*/
|
|
void blk_set_stacking_limits(struct queue_limits *lim)
|
|
{
|
|
blk_set_default_limits(lim);
|
|
|
|
/* Inherit limits from component devices */
|
|
lim->discard_zeroes_data = 1;
|
|
lim->max_segments = USHRT_MAX;
|
|
lim->max_hw_sectors = UINT_MAX;
|
|
lim->max_segment_size = UINT_MAX;
|
|
lim->max_sectors = UINT_MAX;
|
|
lim->max_dev_sectors = UINT_MAX;
|
|
lim->max_write_same_sectors = UINT_MAX;
|
|
}
|
|
EXPORT_SYMBOL(blk_set_stacking_limits);
|
|
|
|
/**
|
|
* blk_queue_make_request - define an alternate make_request function for a device
|
|
* @q: the request queue for the device to be affected
|
|
* @mfn: the alternate make_request function
|
|
*
|
|
* Description:
|
|
* The normal way for &struct bios to be passed to a device
|
|
* driver is for them to be collected into requests on a request
|
|
* queue, and then to allow the device driver to select requests
|
|
* off that queue when it is ready. This works well for many block
|
|
* devices. However some block devices (typically virtual devices
|
|
* such as md or lvm) do not benefit from the processing on the
|
|
* request queue, and are served best by having the requests passed
|
|
* directly to them. This can be achieved by providing a function
|
|
* to blk_queue_make_request().
|
|
*
|
|
* Caveat:
|
|
* The driver that does this *must* be able to deal appropriately
|
|
* with buffers in "highmemory". This can be accomplished by either calling
|
|
* __bio_kmap_atomic() to get a temporary kernel mapping, or by calling
|
|
* blk_queue_bounce() to create a buffer in normal memory.
|
|
**/
|
|
void blk_queue_make_request(struct request_queue *q, make_request_fn *mfn)
|
|
{
|
|
/*
|
|
* set defaults
|
|
*/
|
|
q->nr_requests = BLKDEV_MAX_RQ;
|
|
|
|
q->make_request_fn = mfn;
|
|
blk_queue_dma_alignment(q, 511);
|
|
blk_queue_congestion_threshold(q);
|
|
q->nr_batching = BLK_BATCH_REQ;
|
|
|
|
blk_set_default_limits(&q->limits);
|
|
|
|
/*
|
|
* by default assume old behaviour and bounce for any highmem page
|
|
*/
|
|
blk_queue_bounce_limit(q, BLK_BOUNCE_HIGH);
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_make_request);
|
|
|
|
/**
|
|
* blk_queue_bounce_limit - set bounce buffer limit for queue
|
|
* @q: the request queue for the device
|
|
* @max_addr: the maximum address the device can handle
|
|
*
|
|
* Description:
|
|
* Different hardware can have different requirements as to what pages
|
|
* it can do I/O directly to. A low level driver can call
|
|
* blk_queue_bounce_limit to have lower memory pages allocated as bounce
|
|
* buffers for doing I/O to pages residing above @max_addr.
|
|
**/
|
|
void blk_queue_bounce_limit(struct request_queue *q, u64 max_addr)
|
|
{
|
|
unsigned long b_pfn = max_addr >> PAGE_SHIFT;
|
|
int dma = 0;
|
|
|
|
q->bounce_gfp = GFP_NOIO;
|
|
#if BITS_PER_LONG == 64
|
|
/*
|
|
* Assume anything <= 4GB can be handled by IOMMU. Actually
|
|
* some IOMMUs can handle everything, but I don't know of a
|
|
* way to test this here.
|
|
*/
|
|
if (b_pfn < (min_t(u64, 0xffffffffUL, BLK_BOUNCE_HIGH) >> PAGE_SHIFT))
|
|
dma = 1;
|
|
q->limits.bounce_pfn = max(max_low_pfn, b_pfn);
|
|
#else
|
|
if (b_pfn < blk_max_low_pfn)
|
|
dma = 1;
|
|
q->limits.bounce_pfn = b_pfn;
|
|
#endif
|
|
if (dma) {
|
|
init_emergency_isa_pool();
|
|
q->bounce_gfp = GFP_NOIO | GFP_DMA;
|
|
q->limits.bounce_pfn = b_pfn;
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_bounce_limit);
|
|
|
|
/**
|
|
* blk_queue_max_hw_sectors - set max sectors for a request for this queue
|
|
* @q: the request queue for the device
|
|
* @max_hw_sectors: max hardware sectors in the usual 512b unit
|
|
*
|
|
* Description:
|
|
* Enables a low level driver to set a hard upper limit,
|
|
* max_hw_sectors, on the size of requests. max_hw_sectors is set by
|
|
* the device driver based upon the capabilities of the I/O
|
|
* controller.
|
|
*
|
|
* max_dev_sectors is a hard limit imposed by the storage device for
|
|
* READ/WRITE requests. It is set by the disk driver.
|
|
*
|
|
* max_sectors is a soft limit imposed by the block layer for
|
|
* filesystem type requests. This value can be overridden on a
|
|
* per-device basis in /sys/block/<device>/queue/max_sectors_kb.
|
|
* The soft limit can not exceed max_hw_sectors.
|
|
**/
|
|
void blk_queue_max_hw_sectors(struct request_queue *q, unsigned int max_hw_sectors)
|
|
{
|
|
struct queue_limits *limits = &q->limits;
|
|
unsigned int max_sectors;
|
|
|
|
if ((max_hw_sectors << 9) < PAGE_CACHE_SIZE) {
|
|
max_hw_sectors = 1 << (PAGE_CACHE_SHIFT - 9);
|
|
printk(KERN_INFO "%s: set to minimum %d\n",
|
|
__func__, max_hw_sectors);
|
|
}
|
|
|
|
limits->max_hw_sectors = max_hw_sectors;
|
|
max_sectors = min_not_zero(max_hw_sectors, limits->max_dev_sectors);
|
|
max_sectors = min_t(unsigned int, max_sectors, BLK_DEF_MAX_SECTORS);
|
|
limits->max_sectors = max_sectors;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_max_hw_sectors);
|
|
|
|
/**
|
|
* blk_queue_chunk_sectors - set size of the chunk for this queue
|
|
* @q: the request queue for the device
|
|
* @chunk_sectors: chunk sectors in the usual 512b unit
|
|
*
|
|
* Description:
|
|
* If a driver doesn't want IOs to cross a given chunk size, it can set
|
|
* this limit and prevent merging across chunks. Note that the chunk size
|
|
* must currently be a power-of-2 in sectors. Also note that the block
|
|
* layer must accept a page worth of data at any offset. So if the
|
|
* crossing of chunks is a hard limitation in the driver, it must still be
|
|
* prepared to split single page bios.
|
|
**/
|
|
void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors)
|
|
{
|
|
BUG_ON(!is_power_of_2(chunk_sectors));
|
|
q->limits.chunk_sectors = chunk_sectors;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_chunk_sectors);
|
|
|
|
/**
|
|
* blk_queue_max_discard_sectors - set max sectors for a single discard
|
|
* @q: the request queue for the device
|
|
* @max_discard_sectors: maximum number of sectors to discard
|
|
**/
|
|
void blk_queue_max_discard_sectors(struct request_queue *q,
|
|
unsigned int max_discard_sectors)
|
|
{
|
|
q->limits.max_hw_discard_sectors = max_discard_sectors;
|
|
q->limits.max_discard_sectors = max_discard_sectors;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_max_discard_sectors);
|
|
|
|
/**
|
|
* blk_queue_max_write_same_sectors - set max sectors for a single write same
|
|
* @q: the request queue for the device
|
|
* @max_write_same_sectors: maximum number of sectors to write per command
|
|
**/
|
|
void blk_queue_max_write_same_sectors(struct request_queue *q,
|
|
unsigned int max_write_same_sectors)
|
|
{
|
|
q->limits.max_write_same_sectors = max_write_same_sectors;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_max_write_same_sectors);
|
|
|
|
/**
|
|
* blk_queue_max_segments - set max hw segments for a request for this queue
|
|
* @q: the request queue for the device
|
|
* @max_segments: max number of segments
|
|
*
|
|
* Description:
|
|
* Enables a low level driver to set an upper limit on the number of
|
|
* hw data segments in a request.
|
|
**/
|
|
void blk_queue_max_segments(struct request_queue *q, unsigned short max_segments)
|
|
{
|
|
if (!max_segments) {
|
|
max_segments = 1;
|
|
printk(KERN_INFO "%s: set to minimum %d\n",
|
|
__func__, max_segments);
|
|
}
|
|
|
|
q->limits.max_segments = max_segments;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_max_segments);
|
|
|
|
/**
|
|
* blk_queue_max_segment_size - set max segment size for blk_rq_map_sg
|
|
* @q: the request queue for the device
|
|
* @max_size: max size of segment in bytes
|
|
*
|
|
* Description:
|
|
* Enables a low level driver to set an upper limit on the size of a
|
|
* coalesced segment
|
|
**/
|
|
void blk_queue_max_segment_size(struct request_queue *q, unsigned int max_size)
|
|
{
|
|
if (max_size < PAGE_CACHE_SIZE) {
|
|
max_size = PAGE_CACHE_SIZE;
|
|
printk(KERN_INFO "%s: set to minimum %d\n",
|
|
__func__, max_size);
|
|
}
|
|
|
|
q->limits.max_segment_size = max_size;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_max_segment_size);
|
|
|
|
/**
|
|
* blk_queue_logical_block_size - set logical block size for the queue
|
|
* @q: the request queue for the device
|
|
* @size: the logical block size, in bytes
|
|
*
|
|
* Description:
|
|
* This should be set to the lowest possible block size that the
|
|
* storage device can address. The default of 512 covers most
|
|
* hardware.
|
|
**/
|
|
void blk_queue_logical_block_size(struct request_queue *q, unsigned short size)
|
|
{
|
|
q->limits.logical_block_size = size;
|
|
|
|
if (q->limits.physical_block_size < size)
|
|
q->limits.physical_block_size = size;
|
|
|
|
if (q->limits.io_min < q->limits.physical_block_size)
|
|
q->limits.io_min = q->limits.physical_block_size;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_logical_block_size);
|
|
|
|
/**
|
|
* blk_queue_physical_block_size - set physical block size for the queue
|
|
* @q: the request queue for the device
|
|
* @size: the physical block size, in bytes
|
|
*
|
|
* Description:
|
|
* This should be set to the lowest possible sector size that the
|
|
* hardware can operate on without reverting to read-modify-write
|
|
* operations.
|
|
*/
|
|
void blk_queue_physical_block_size(struct request_queue *q, unsigned int size)
|
|
{
|
|
q->limits.physical_block_size = size;
|
|
|
|
if (q->limits.physical_block_size < q->limits.logical_block_size)
|
|
q->limits.physical_block_size = q->limits.logical_block_size;
|
|
|
|
if (q->limits.io_min < q->limits.physical_block_size)
|
|
q->limits.io_min = q->limits.physical_block_size;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_physical_block_size);
|
|
|
|
/**
|
|
* blk_queue_alignment_offset - set physical block alignment offset
|
|
* @q: the request queue for the device
|
|
* @offset: alignment offset in bytes
|
|
*
|
|
* Description:
|
|
* Some devices are naturally misaligned to compensate for things like
|
|
* the legacy DOS partition table 63-sector offset. Low-level drivers
|
|
* should call this function for devices whose first sector is not
|
|
* naturally aligned.
|
|
*/
|
|
void blk_queue_alignment_offset(struct request_queue *q, unsigned int offset)
|
|
{
|
|
q->limits.alignment_offset =
|
|
offset & (q->limits.physical_block_size - 1);
|
|
q->limits.misaligned = 0;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_alignment_offset);
|
|
|
|
/**
|
|
* blk_limits_io_min - set minimum request size for a device
|
|
* @limits: the queue limits
|
|
* @min: smallest I/O size in bytes
|
|
*
|
|
* Description:
|
|
* Some devices have an internal block size bigger than the reported
|
|
* hardware sector size. This function can be used to signal the
|
|
* smallest I/O the device can perform without incurring a performance
|
|
* penalty.
|
|
*/
|
|
void blk_limits_io_min(struct queue_limits *limits, unsigned int min)
|
|
{
|
|
limits->io_min = min;
|
|
|
|
if (limits->io_min < limits->logical_block_size)
|
|
limits->io_min = limits->logical_block_size;
|
|
|
|
if (limits->io_min < limits->physical_block_size)
|
|
limits->io_min = limits->physical_block_size;
|
|
}
|
|
EXPORT_SYMBOL(blk_limits_io_min);
|
|
|
|
/**
|
|
* blk_queue_io_min - set minimum request size for the queue
|
|
* @q: the request queue for the device
|
|
* @min: smallest I/O size in bytes
|
|
*
|
|
* Description:
|
|
* Storage devices may report a granularity or preferred minimum I/O
|
|
* size which is the smallest request the device can perform without
|
|
* incurring a performance penalty. For disk drives this is often the
|
|
* physical block size. For RAID arrays it is often the stripe chunk
|
|
* size. A properly aligned multiple of minimum_io_size is the
|
|
* preferred request size for workloads where a high number of I/O
|
|
* operations is desired.
|
|
*/
|
|
void blk_queue_io_min(struct request_queue *q, unsigned int min)
|
|
{
|
|
blk_limits_io_min(&q->limits, min);
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_io_min);
|
|
|
|
/**
|
|
* blk_limits_io_opt - set optimal request size for a device
|
|
* @limits: the queue limits
|
|
* @opt: smallest I/O size in bytes
|
|
*
|
|
* Description:
|
|
* Storage devices may report an optimal I/O size, which is the
|
|
* device's preferred unit for sustained I/O. This is rarely reported
|
|
* for disk drives. For RAID arrays it is usually the stripe width or
|
|
* the internal track size. A properly aligned multiple of
|
|
* optimal_io_size is the preferred request size for workloads where
|
|
* sustained throughput is desired.
|
|
*/
|
|
void blk_limits_io_opt(struct queue_limits *limits, unsigned int opt)
|
|
{
|
|
limits->io_opt = opt;
|
|
}
|
|
EXPORT_SYMBOL(blk_limits_io_opt);
|
|
|
|
/**
|
|
* blk_queue_io_opt - set optimal request size for the queue
|
|
* @q: the request queue for the device
|
|
* @opt: optimal request size in bytes
|
|
*
|
|
* Description:
|
|
* Storage devices may report an optimal I/O size, which is the
|
|
* device's preferred unit for sustained I/O. This is rarely reported
|
|
* for disk drives. For RAID arrays it is usually the stripe width or
|
|
* the internal track size. A properly aligned multiple of
|
|
* optimal_io_size is the preferred request size for workloads where
|
|
* sustained throughput is desired.
|
|
*/
|
|
void blk_queue_io_opt(struct request_queue *q, unsigned int opt)
|
|
{
|
|
blk_limits_io_opt(&q->limits, opt);
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_io_opt);
|
|
|
|
/**
|
|
* blk_queue_stack_limits - inherit underlying queue limits for stacked drivers
|
|
* @t: the stacking driver (top)
|
|
* @b: the underlying device (bottom)
|
|
**/
|
|
void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b)
|
|
{
|
|
blk_stack_limits(&t->limits, &b->limits, 0);
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_stack_limits);
|
|
|
|
/**
|
|
* blk_stack_limits - adjust queue_limits for stacked devices
|
|
* @t: the stacking driver limits (top device)
|
|
* @b: the underlying queue limits (bottom, component device)
|
|
* @start: first data sector within component device
|
|
*
|
|
* Description:
|
|
* This function is used by stacking drivers like MD and DM to ensure
|
|
* that all component devices have compatible block sizes and
|
|
* alignments. The stacking driver must provide a queue_limits
|
|
* struct (top) and then iteratively call the stacking function for
|
|
* all component (bottom) devices. The stacking function will
|
|
* attempt to combine the values and ensure proper alignment.
|
|
*
|
|
* Returns 0 if the top and bottom queue_limits are compatible. The
|
|
* top device's block sizes and alignment offsets may be adjusted to
|
|
* ensure alignment with the bottom device. If no compatible sizes
|
|
* and alignments exist, -1 is returned and the resulting top
|
|
* queue_limits will have the misaligned flag set to indicate that
|
|
* the alignment_offset is undefined.
|
|
*/
|
|
int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
|
|
sector_t start)
|
|
{
|
|
unsigned int top, bottom, alignment, ret = 0;
|
|
|
|
t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors);
|
|
t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors);
|
|
t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors);
|
|
t->max_write_same_sectors = min(t->max_write_same_sectors,
|
|
b->max_write_same_sectors);
|
|
t->bounce_pfn = min_not_zero(t->bounce_pfn, b->bounce_pfn);
|
|
|
|
t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask,
|
|
b->seg_boundary_mask);
|
|
t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask,
|
|
b->virt_boundary_mask);
|
|
|
|
t->max_segments = min_not_zero(t->max_segments, b->max_segments);
|
|
t->max_integrity_segments = min_not_zero(t->max_integrity_segments,
|
|
b->max_integrity_segments);
|
|
|
|
t->max_segment_size = min_not_zero(t->max_segment_size,
|
|
b->max_segment_size);
|
|
|
|
t->misaligned |= b->misaligned;
|
|
|
|
alignment = queue_limit_alignment_offset(b, start);
|
|
|
|
/* Bottom device has different alignment. Check that it is
|
|
* compatible with the current top alignment.
|
|
*/
|
|
if (t->alignment_offset != alignment) {
|
|
|
|
top = max(t->physical_block_size, t->io_min)
|
|
+ t->alignment_offset;
|
|
bottom = max(b->physical_block_size, b->io_min) + alignment;
|
|
|
|
/* Verify that top and bottom intervals line up */
|
|
if (max(top, bottom) % min(top, bottom)) {
|
|
t->misaligned = 1;
|
|
ret = -1;
|
|
}
|
|
}
|
|
|
|
t->logical_block_size = max(t->logical_block_size,
|
|
b->logical_block_size);
|
|
|
|
t->physical_block_size = max(t->physical_block_size,
|
|
b->physical_block_size);
|
|
|
|
t->io_min = max(t->io_min, b->io_min);
|
|
t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
|
|
|
|
t->cluster &= b->cluster;
|
|
t->discard_zeroes_data &= b->discard_zeroes_data;
|
|
|
|
/* Physical block size a multiple of the logical block size? */
|
|
if (t->physical_block_size & (t->logical_block_size - 1)) {
|
|
t->physical_block_size = t->logical_block_size;
|
|
t->misaligned = 1;
|
|
ret = -1;
|
|
}
|
|
|
|
/* Minimum I/O a multiple of the physical block size? */
|
|
if (t->io_min & (t->physical_block_size - 1)) {
|
|
t->io_min = t->physical_block_size;
|
|
t->misaligned = 1;
|
|
ret = -1;
|
|
}
|
|
|
|
/* Optimal I/O a multiple of the physical block size? */
|
|
if (t->io_opt & (t->physical_block_size - 1)) {
|
|
t->io_opt = 0;
|
|
t->misaligned = 1;
|
|
ret = -1;
|
|
}
|
|
|
|
t->raid_partial_stripes_expensive =
|
|
max(t->raid_partial_stripes_expensive,
|
|
b->raid_partial_stripes_expensive);
|
|
|
|
/* Find lowest common alignment_offset */
|
|
t->alignment_offset = lcm_not_zero(t->alignment_offset, alignment)
|
|
% max(t->physical_block_size, t->io_min);
|
|
|
|
/* Verify that new alignment_offset is on a logical block boundary */
|
|
if (t->alignment_offset & (t->logical_block_size - 1)) {
|
|
t->misaligned = 1;
|
|
ret = -1;
|
|
}
|
|
|
|
/* Discard alignment and granularity */
|
|
if (b->discard_granularity) {
|
|
alignment = queue_limit_discard_alignment(b, start);
|
|
|
|
if (t->discard_granularity != 0 &&
|
|
t->discard_alignment != alignment) {
|
|
top = t->discard_granularity + t->discard_alignment;
|
|
bottom = b->discard_granularity + alignment;
|
|
|
|
/* Verify that top and bottom intervals line up */
|
|
if ((max(top, bottom) % min(top, bottom)) != 0)
|
|
t->discard_misaligned = 1;
|
|
}
|
|
|
|
t->max_discard_sectors = min_not_zero(t->max_discard_sectors,
|
|
b->max_discard_sectors);
|
|
t->max_hw_discard_sectors = min_not_zero(t->max_hw_discard_sectors,
|
|
b->max_hw_discard_sectors);
|
|
t->discard_granularity = max(t->discard_granularity,
|
|
b->discard_granularity);
|
|
t->discard_alignment = lcm_not_zero(t->discard_alignment, alignment) %
|
|
t->discard_granularity;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(blk_stack_limits);
|
|
|
|
/**
|
|
* bdev_stack_limits - adjust queue limits for stacked drivers
|
|
* @t: the stacking driver limits (top device)
|
|
* @bdev: the component block_device (bottom)
|
|
* @start: first data sector within component device
|
|
*
|
|
* Description:
|
|
* Merges queue limits for a top device and a block_device. Returns
|
|
* 0 if alignment didn't change. Returns -1 if adding the bottom
|
|
* device caused misalignment.
|
|
*/
|
|
int bdev_stack_limits(struct queue_limits *t, struct block_device *bdev,
|
|
sector_t start)
|
|
{
|
|
struct request_queue *bq = bdev_get_queue(bdev);
|
|
|
|
start += get_start_sect(bdev);
|
|
|
|
return blk_stack_limits(t, &bq->limits, start);
|
|
}
|
|
EXPORT_SYMBOL(bdev_stack_limits);
|
|
|
|
/**
|
|
* disk_stack_limits - adjust queue limits for stacked drivers
|
|
* @disk: MD/DM gendisk (top)
|
|
* @bdev: the underlying block device (bottom)
|
|
* @offset: offset to beginning of data within component device
|
|
*
|
|
* Description:
|
|
* Merges the limits for a top level gendisk and a bottom level
|
|
* block_device.
|
|
*/
|
|
void disk_stack_limits(struct gendisk *disk, struct block_device *bdev,
|
|
sector_t offset)
|
|
{
|
|
struct request_queue *t = disk->queue;
|
|
|
|
if (bdev_stack_limits(&t->limits, bdev, offset >> 9) < 0) {
|
|
char top[BDEVNAME_SIZE], bottom[BDEVNAME_SIZE];
|
|
|
|
disk_name(disk, 0, top);
|
|
bdevname(bdev, bottom);
|
|
|
|
printk(KERN_NOTICE "%s: Warning: Device %s is misaligned\n",
|
|
top, bottom);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(disk_stack_limits);
|
|
|
|
/**
|
|
* blk_queue_dma_pad - set pad mask
|
|
* @q: the request queue for the device
|
|
* @mask: pad mask
|
|
*
|
|
* Set dma pad mask.
|
|
*
|
|
* Appending pad buffer to a request modifies the last entry of a
|
|
* scatter list such that it includes the pad buffer.
|
|
**/
|
|
void blk_queue_dma_pad(struct request_queue *q, unsigned int mask)
|
|
{
|
|
q->dma_pad_mask = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_dma_pad);
|
|
|
|
/**
|
|
* blk_queue_update_dma_pad - update pad mask
|
|
* @q: the request queue for the device
|
|
* @mask: pad mask
|
|
*
|
|
* Update dma pad mask.
|
|
*
|
|
* Appending pad buffer to a request modifies the last entry of a
|
|
* scatter list such that it includes the pad buffer.
|
|
**/
|
|
void blk_queue_update_dma_pad(struct request_queue *q, unsigned int mask)
|
|
{
|
|
if (mask > q->dma_pad_mask)
|
|
q->dma_pad_mask = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_update_dma_pad);
|
|
|
|
/**
|
|
* blk_queue_dma_drain - Set up a drain buffer for excess dma.
|
|
* @q: the request queue for the device
|
|
* @dma_drain_needed: fn which returns non-zero if drain is necessary
|
|
* @buf: physically contiguous buffer
|
|
* @size: size of the buffer in bytes
|
|
*
|
|
* Some devices have excess DMA problems and can't simply discard (or
|
|
* zero fill) the unwanted piece of the transfer. They have to have a
|
|
* real area of memory to transfer it into. The use case for this is
|
|
* ATAPI devices in DMA mode. If the packet command causes a transfer
|
|
* bigger than the transfer size some HBAs will lock up if there
|
|
* aren't DMA elements to contain the excess transfer. What this API
|
|
* does is adjust the queue so that the buf is always appended
|
|
* silently to the scatterlist.
|
|
*
|
|
* Note: This routine adjusts max_hw_segments to make room for appending
|
|
* the drain buffer. If you call blk_queue_max_segments() after calling
|
|
* this routine, you must set the limit to one fewer than your device
|
|
* can support otherwise there won't be room for the drain buffer.
|
|
*/
|
|
int blk_queue_dma_drain(struct request_queue *q,
|
|
dma_drain_needed_fn *dma_drain_needed,
|
|
void *buf, unsigned int size)
|
|
{
|
|
if (queue_max_segments(q) < 2)
|
|
return -EINVAL;
|
|
/* make room for appending the drain */
|
|
blk_queue_max_segments(q, queue_max_segments(q) - 1);
|
|
q->dma_drain_needed = dma_drain_needed;
|
|
q->dma_drain_buffer = buf;
|
|
q->dma_drain_size = size;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_dma_drain);
|
|
|
|
/**
|
|
* blk_queue_segment_boundary - set boundary rules for segment merging
|
|
* @q: the request queue for the device
|
|
* @mask: the memory boundary mask
|
|
**/
|
|
void blk_queue_segment_boundary(struct request_queue *q, unsigned long mask)
|
|
{
|
|
if (mask < PAGE_CACHE_SIZE - 1) {
|
|
mask = PAGE_CACHE_SIZE - 1;
|
|
printk(KERN_INFO "%s: set to minimum %lx\n",
|
|
__func__, mask);
|
|
}
|
|
|
|
q->limits.seg_boundary_mask = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_segment_boundary);
|
|
|
|
/**
|
|
* blk_queue_virt_boundary - set boundary rules for bio merging
|
|
* @q: the request queue for the device
|
|
* @mask: the memory boundary mask
|
|
**/
|
|
void blk_queue_virt_boundary(struct request_queue *q, unsigned long mask)
|
|
{
|
|
q->limits.virt_boundary_mask = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_virt_boundary);
|
|
|
|
/**
|
|
* blk_queue_dma_alignment - set dma length and memory alignment
|
|
* @q: the request queue for the device
|
|
* @mask: alignment mask
|
|
*
|
|
* description:
|
|
* set required memory and length alignment for direct dma transactions.
|
|
* this is used when building direct io requests for the queue.
|
|
*
|
|
**/
|
|
void blk_queue_dma_alignment(struct request_queue *q, int mask)
|
|
{
|
|
q->dma_alignment = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_dma_alignment);
|
|
|
|
/**
|
|
* blk_queue_update_dma_alignment - update dma length and memory alignment
|
|
* @q: the request queue for the device
|
|
* @mask: alignment mask
|
|
*
|
|
* description:
|
|
* update required memory and length alignment for direct dma transactions.
|
|
* If the requested alignment is larger than the current alignment, then
|
|
* the current queue alignment is updated to the new value, otherwise it
|
|
* is left alone. The design of this is to allow multiple objects
|
|
* (driver, device, transport etc) to set their respective
|
|
* alignments without having them interfere.
|
|
*
|
|
**/
|
|
void blk_queue_update_dma_alignment(struct request_queue *q, int mask)
|
|
{
|
|
BUG_ON(mask > PAGE_SIZE);
|
|
|
|
if (mask > q->dma_alignment)
|
|
q->dma_alignment = mask;
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_update_dma_alignment);
|
|
|
|
/**
|
|
* blk_queue_flush - configure queue's cache flush capability
|
|
* @q: the request queue for the device
|
|
* @flush: 0, REQ_FLUSH or REQ_FLUSH | REQ_FUA | REQ_BARRIER
|
|
*
|
|
* Tell block layer cache flush capability of @q. If it supports
|
|
* flushing, REQ_FLUSH should be set. If it supports bypassing
|
|
* write cache for individual writes, REQ_FUA should be set. If cache
|
|
* barrier is supported set REQ_BARRIER.
|
|
*/
|
|
void blk_queue_flush(struct request_queue *q, unsigned int flush)
|
|
{
|
|
WARN_ON_ONCE(flush & ~(REQ_FLUSH | REQ_FUA | REQ_BARRIER));
|
|
|
|
if (WARN_ON_ONCE(!(flush & REQ_FLUSH) && ((flush & REQ_FUA) ||
|
|
(flush & REQ_BARRIER)))) {
|
|
flush &= ~REQ_FUA;
|
|
flush &= ~REQ_BARRIER;
|
|
}
|
|
|
|
q->flush_flags = flush & (REQ_FLUSH | REQ_FUA | REQ_BARRIER);
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_flush);
|
|
|
|
void blk_queue_flush_queueable(struct request_queue *q, bool queueable)
|
|
{
|
|
q->flush_not_queueable = !queueable;
|
|
}
|
|
EXPORT_SYMBOL_GPL(blk_queue_flush_queueable);
|
|
|
|
static int __init blk_settings_init(void)
|
|
{
|
|
blk_max_low_pfn = max_low_pfn - 1;
|
|
blk_max_pfn = max_pfn - 1;
|
|
return 0;
|
|
}
|
|
subsys_initcall(blk_settings_init);
|