Merge 6.6.96 into android15-6.6-lts
GKI (arm64) relevant 31 out of 139 changes, affecting 36 files +289/-14835cda8619d
mailbox: Not protect module_put with spin_lock_irqsave [1 file, +1/-1]a13b2634dd
leds: multicolor: Fix intensity setting while SW blinking [1 file, +2/-1]96715eb1a1
fuse: fix race between concurrent setattrs from multiple nodes [1 file, +11/-0]969c9646f0
PCI: dwc: Make link training more robust by setting PORT_LOGIC_LINK_WIDTH to one lane [1 file, +1/-4]433cb3e70d
usb: Add checks for snprintf() calls in usb_alloc_dev() [1 file, +10/-4]8851e40587
usb: typec: displayport: Receive DP Status Update NAK request exit dp altmode [1 file, +4/-0]23fe269bfa
usb: typec: mux: do not return on EOPNOTSUPP in {mux, switch}_set [1 file, +2/-2]f3465bb320
ALSA: usb-audio: Add a quirk for Lenovo Thinkpad Thunderbolt 3 dock [1 file, +2/-0]d120737a6d
scsi: ufs: core: Don't perform UFS clkscaling during host async scan [1 file, +3/-0]2cbeb47ea9
ovl: Check for NULL d_inode() in ovl_dentry_upper() [1 file, +3/-1]3bc5abf8e7
f2fs: don't over-report free space or inodes in statvfs [1 file, +18/-12]b7904e2fcc
af_unix: Define locking order for unix_table_double_lock(). [1 file, +11/-1]fd3af69c14
af_unix: Define locking order for U_LOCK_SECOND in unix_state_double_lock(). [1 file, +15/-1]0919388ac4
af_unix: Define locking order for U_RECVQ_LOCK_EMBRYO in unix_collect_skb(). [2 files, +22/-7]aabb458c33
af_unix: Don't call skb_get() for OOB skb. [2 files, +7/-36]fad0a2c160
af_unix: Don't leave consecutive consumed OOB skbs. [1 file, +11/-2]64a99eff8d
lib/group_cpus: fix NULL pointer dereference from group_cpus_evenly() [1 file, +8/-1]74fcb3852a
ALSA: usb-audio: Fix out-of-bounds read in snd_usb_get_audioformat_uac3() [1 file, +2/-0]591f796257
attach_recursive_mnt(): do not lock the covering tree when sliding something under it [1 file, +4/-4]d58343f813
af_unix: Don't set -ECONNRESET for consumed OOB skb. [1 file, +12/-6]8f96a2ae16
vsock/uapi: fix linux/vm_sockets.h userspace compilation errors [1 file, +4/-0]0804d77fd6
net: selftests: fix TCP packet checksum [1 file, +3/-2]e395e34b06
Bluetooth: L2CAP: Fix L2CAP MTU negotiation [1 file, +8/-1]33acbeeb5a
HID: wacom: fix memory leak on kobject creation failure [1 file, +3/-1]f531651c39
HID: wacom: fix memory leak on sysfs attribute creation failure [1 file, +1/-0]42cf4f0161
HID: wacom: fix kobject reference count leak [1 file, +1/-0]56aa7679c9
media: uvcvideo: Rollback non processed entities on error [1 file, +26/-13]c46358d027
Revert "ipv6: save dontfrag in cork" [2 files, +4/-6]6f27bbf3de
nvme: always punt polled uring_cmd end_io work to task_work [1 file, +7/-9]67a50f5721
firmware: arm_scmi: Add a common helper to check if a message is supported [2 files, +38/-0]f9917821c5
firmware: arm_scmi: Ensure that the message-id supports fastchannel [2 files, +45/-33] Changes in 6.6.96 cifs: Correctly set SMB1 SessionKey field in Session Setup Request cifs: Fix cifs_query_path_info() for Windows NT servers cifs: Fix encoding of SMB1 Session Setup NTLMSSP Request in non-UNICODE mode NFSv4: Always set NLINK even if the server doesn't support it NFSv4.2: fix listxattr to return selinux security label mailbox: Not protect module_put with spin_lock_irqsave mfd: max14577: Fix wakeup source leaks on device unbind sunrpc: don't immediately retransmit on seqno miss leds: multicolor: Fix intensity setting while SW blinking fuse: fix race between concurrent setattrs from multiple nodes cxl/region: Add a dev_err() on missing target list entries NFSv4: xattr handlers should check for absent nfs filehandles hwmon: (pmbus/max34440) Fix support for max34451 ksmbd: allow a filename to contain special characters on SMB3.1.1 posix extension ksmbd: provide zero as a unique ID to the Mac client rust: module: place cleanup_module() in .exit.text section dmaengine: idxd: Check availability of workqueue allocated by idxd wq driver before using dmaengine: xilinx_dma: Set dma_device directions PCI: dwc: Make link training more robust by setting PORT_LOGIC_LINK_WIDTH to one lane PCI: apple: Fix missing OF node reference in apple_pcie_setup_port md/md-bitmap: fix dm-raid max_write_behind setting amd/amdkfd: fix a kfd_process ref leak bcache: fix NULL pointer in cache_set_flush() drm/scheduler: signal scheduled fence when kill job iio: pressure: zpa2326: Use aligned_s64 for the timestamp um: Add cmpxchg8b_emu and checksum functions to asm-prototypes.h um: use proper care when taking mmap lock during segfault coresight: Only check bottom two claim bits usb: dwc2: also exit clock_gating when stopping udc while suspended iio: adc: ad_sigma_delta: Fix use of uninitialized status_pos misc: tps6594-pfsm: Add NULL pointer check in tps6594_pfsm_probe() usb: potential integer overflow in usbg_make_tpg() tty: serial: uartlite: register uart driver in init usb: common: usb-conn-gpio: use a unique name for usb connector device usb: Add checks for snprintf() calls in usb_alloc_dev() usb: cdc-wdm: avoid setting WDM_READ for ZLP-s usb: typec: displayport: Receive DP Status Update NAK request exit dp altmode usb: typec: mux: do not return on EOPNOTSUPP in {mux, switch}_set ALSA: hda: Ignore unsol events for cards being shut down ALSA: hda: Add new pci id for AMD GPU display HD audio controller ALSA: usb-audio: Add a quirk for Lenovo Thinkpad Thunderbolt 3 dock ceph: fix possible integer overflow in ceph_zero_objects() scsi: ufs: core: Don't perform UFS clkscaling during host async scan ovl: Check for NULL d_inode() in ovl_dentry_upper() btrfs: handle csum tree error with rescue=ibadroots correctly fs/jfs: consolidate sanity checking in dbMount jfs: validate AG parameters in dbMount() to prevent crashes ASoC: codecs: wcd9335: Handle nicer probe deferral and simplify with dev_err_probe() ASoC: codec: wcd9335: Convert to GPIO descriptors ASoC: codecs: wcd9335: Fix missing free of regulator supplies f2fs: don't over-report free space or inodes in statvfs Drivers: hv: vmbus: Add utility function for querying ring size uio_hv_generic: Query the ringbuffer size for device uio_hv_generic: Align ring size to system page PCI: apple: Use helper function for_each_child_of_node_scoped() PCI: apple: Set only available ports up tty: vt: make init parameter of consw::con_init() a bool tty: vt: sanitize arguments of consw::con_clear() tty: vt: make consw::con_switch() return a bool dummycon: Trigger redraw when switching consoles with deferred takeover platform/x86: ideapad-laptop: introduce a generic notification chain platform/x86: ideapad-laptop: move ymc_trigger_ec from lenovo-ymc platform/x86: ideapad-laptop: move ACPI helpers from header to source file platform/x86: ideapad-laptop: use usleep_range() for EC polling af_unix: Define locking order for unix_table_double_lock(). af_unix: Define locking order for U_LOCK_SECOND in unix_state_double_lock(). af_unix: Define locking order for U_RECVQ_LOCK_EMBRYO in unix_collect_skb(). af_unix: Don't call skb_get() for OOB skb. af_unix: Don't leave consecutive consumed OOB skbs. i2c: tiny-usb: disable zero-length read messages i2c: robotfuzz-osif: disable zero-length read messages mm/damon/sysfs-schemes: free old damon_sysfs_scheme_filter->memcg_path on write ASoC: amd: yc: Add DMI quirk for Lenovo IdeaPad Slim 5 15 s390/pkey: Prevent overflow in size calculation for memdup_user() lib/group_cpus: fix NULL pointer dereference from group_cpus_evenly() atm: clip: prevent NULL deref in clip_push() ALSA: usb-audio: Fix out-of-bounds read in snd_usb_get_audioformat_uac3() attach_recursive_mnt(): do not lock the covering tree when sliding something under it libbpf: Fix null pointer dereference in btf_dump__free on allocation failure wifi: mac80211: fix beacon interval calculation overflow af_unix: Don't set -ECONNRESET for consumed OOB skb. vsock/uapi: fix linux/vm_sockets.h userspace compilation errors um: ubd: Add missing error check in start_io_thread() libbpf: Fix possible use-after-free for externs net: enetc: Correct endianness handling in _enetc_rd_reg64 atm: Release atm_dev_mutex after removing procfs in atm_dev_deregister(). ALSA: hda/realtek: Fix built-in mic on ASUS VivoBook X507UAR net: selftests: fix TCP packet checksum drm/i915: fix build error some more drm/bridge: ti-sn65dsi86: make use of debugfs_init callback drm/bridge: ti-sn65dsi86: Add HPD for DisplayPort connector type smb: client: fix potential deadlock when reconnecting channels EDAC/amd64: Fix size calculation for Non-Power-of-Two DIMMs staging: rtl8723bs: Avoid memset() in aes_cipher() and aes_decipher() dt-bindings: serial: 8250: Make clocks and clock-frequency exclusive serial: imx: Restore original RXTL for console to fix data loss Bluetooth: L2CAP: Fix L2CAP MTU negotiation dm-raid: fix variable in journal device check btrfs: fix a race between renames and directory logging btrfs: update superblock's device bytes_used when dropping chunk net: libwx: fix the creation of page_pool HID: lenovo: Restrict F7/9/11 mode to compact keyboards only HID: wacom: fix memory leak on kobject creation failure HID: wacom: fix memory leak on sysfs attribute creation failure HID: wacom: fix kobject reference count leak scsi: megaraid_sas: Fix invalid node index drm/ast: Fix comment on modeset lock drm/cirrus-qemu: Fix pitch programming drm/etnaviv: Protect the scheduler's pending list with its lock drm/tegra: Assign plane type before registration drm/tegra: Fix a possible null pointer dereference drm/udl: Unregister device before cleaning up on disconnect drm/msm/gpu: Fix crash when throttling GPU immediately during boot drm/amdkfd: Fix race in GWS queue scheduling drm/bridge: cdns-dsi: Fix the clock variable for mode_valid() drm/bridge: cdns-dsi: Fix phy de-init and flag it so drm/bridge: cdns-dsi: Fix connecting to next bridge drm/bridge: cdns-dsi: Check return value when getting default PHY config drm/bridge: cdns-dsi: Wait for Clk and Data Lanes to be ready drm/amd/display: Add null pointer check for get_first_active_display() drm/amdgpu: amdgpu_vram_mgr_new(): Clamp lpfn to total vram drm/i915/gem: Allow EXEC_CAPTURE on recoverable contexts on DG1 drm/amdgpu: Add kicker device detection drm/amdgpu: switch job hw_fence to amdgpu_fence ksmbd: Use unsafe_memcpy() for ntlm_negotiate ksmbd: remove unsafe_memcpy use in session setup scripts: clean up IA-64 code kbuild: rpm-pkg: simplify installkernel %post media: uvcvideo: Rollback non processed entities on error s390/entry: Fix last breaking event handling in case of stack corruption Kunit to check the longest symbol length x86/tools: Drop duplicate unlikely() definition in insn_decoder_test.c Revert "ipv6: save dontfrag in cork" spi: spi-cadence-quadspi: Fix pm runtime unbalance nvme: always punt polled uring_cmd end_io work to task_work firmware: arm_scmi: Add a common helper to check if a message is supported firmware: arm_scmi: Ensure that the message-id supports fastchannel ARM: 9354/1: ptrace: Use bitfield helpers Linux 6.6.96 Change-Id: Iadeae78f7c4d52dcb2037f88c6d651626f62054b Signed-off-by: Greg Kroah-Hartman <gregkh@google.com>
This commit is contained in:
@@ -45,7 +45,7 @@ allOf:
|
|||||||
- ns16550
|
- ns16550
|
||||||
- ns16550a
|
- ns16550a
|
||||||
then:
|
then:
|
||||||
anyOf:
|
oneOf:
|
||||||
- required: [ clock-frequency ]
|
- required: [ clock-frequency ]
|
||||||
- required: [ clocks ]
|
- required: [ clocks ]
|
||||||
|
|
||||||
|
2
Makefile
2
Makefile
@@ -1,7 +1,7 @@
|
|||||||
# SPDX-License-Identifier: GPL-2.0
|
# SPDX-License-Identifier: GPL-2.0
|
||||||
VERSION = 6
|
VERSION = 6
|
||||||
PATCHLEVEL = 6
|
PATCHLEVEL = 6
|
||||||
SUBLEVEL = 95
|
SUBLEVEL = 96
|
||||||
EXTRAVERSION =
|
EXTRAVERSION =
|
||||||
NAME = Pinguïn Aangedreven
|
NAME = Pinguïn Aangedreven
|
||||||
|
|
||||||
|
@@ -10,6 +10,7 @@
|
|||||||
#include <uapi/asm/ptrace.h>
|
#include <uapi/asm/ptrace.h>
|
||||||
|
|
||||||
#ifndef __ASSEMBLY__
|
#ifndef __ASSEMBLY__
|
||||||
|
#include <linux/bitfield.h>
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
|
|
||||||
struct pt_regs {
|
struct pt_regs {
|
||||||
@@ -35,8 +36,8 @@ struct svc_pt_regs {
|
|||||||
|
|
||||||
#ifndef CONFIG_CPU_V7M
|
#ifndef CONFIG_CPU_V7M
|
||||||
#define isa_mode(regs) \
|
#define isa_mode(regs) \
|
||||||
((((regs)->ARM_cpsr & PSR_J_BIT) >> (__ffs(PSR_J_BIT) - 1)) | \
|
(FIELD_GET(PSR_J_BIT, (regs)->ARM_cpsr) << 1 | \
|
||||||
(((regs)->ARM_cpsr & PSR_T_BIT) >> (__ffs(PSR_T_BIT))))
|
FIELD_GET(PSR_T_BIT, (regs)->ARM_cpsr))
|
||||||
#else
|
#else
|
||||||
#define isa_mode(regs) 1 /* Thumb */
|
#define isa_mode(regs) 1 /* Thumb */
|
||||||
#endif
|
#endif
|
||||||
|
@@ -639,7 +639,7 @@ SYM_CODE_START(stack_overflow)
|
|||||||
stmg %r0,%r7,__PT_R0(%r11)
|
stmg %r0,%r7,__PT_R0(%r11)
|
||||||
stmg %r8,%r9,__PT_PSW(%r11)
|
stmg %r8,%r9,__PT_PSW(%r11)
|
||||||
mvc __PT_R8(64,%r11),0(%r14)
|
mvc __PT_R8(64,%r11),0(%r14)
|
||||||
stg %r10,__PT_ORIG_GPR2(%r11) # store last break to orig_gpr2
|
mvc __PT_ORIG_GPR2(8,%r11),__LC_PGM_LAST_BREAK
|
||||||
xc __SF_BACKCHAIN(8,%r15),__SF_BACKCHAIN(%r15)
|
xc __SF_BACKCHAIN(8,%r15),__SF_BACKCHAIN(%r15)
|
||||||
lgr %r2,%r11 # pass pointer to pt_regs
|
lgr %r2,%r11 # pass pointer to pt_regs
|
||||||
jg kernel_stack_overflow
|
jg kernel_stack_overflow
|
||||||
|
@@ -41,7 +41,7 @@ int start_io_thread(unsigned long sp, int *fd_out)
|
|||||||
*fd_out = fds[1];
|
*fd_out = fds[1];
|
||||||
|
|
||||||
err = os_set_fd_block(*fd_out, 0);
|
err = os_set_fd_block(*fd_out, 0);
|
||||||
err = os_set_fd_block(kernel_fd, 0);
|
err |= os_set_fd_block(kernel_fd, 0);
|
||||||
if (err) {
|
if (err) {
|
||||||
printk("start_io_thread - failed to set nonblocking I/O.\n");
|
printk("start_io_thread - failed to set nonblocking I/O.\n");
|
||||||
goto out_close;
|
goto out_close;
|
||||||
|
@@ -1 +1,6 @@
|
|||||||
#include <asm-generic/asm-prototypes.h>
|
#include <asm-generic/asm-prototypes.h>
|
||||||
|
#include <asm/checksum.h>
|
||||||
|
|
||||||
|
#ifdef CONFIG_UML_X86
|
||||||
|
extern void cmpxchg8b_emu(void);
|
||||||
|
#endif
|
||||||
|
@@ -17,6 +17,122 @@
|
|||||||
#include <os.h>
|
#include <os.h>
|
||||||
#include <skas.h>
|
#include <skas.h>
|
||||||
|
|
||||||
|
/*
|
||||||
|
* NOTE: UML does not have exception tables. As such, this is almost a copy
|
||||||
|
* of the code in mm/memory.c, only adjusting the logic to simply check whether
|
||||||
|
* we are coming from the kernel instead of doing an additional lookup in the
|
||||||
|
* exception table.
|
||||||
|
* We can do this simplification because we never get here if the exception was
|
||||||
|
* fixable.
|
||||||
|
*/
|
||||||
|
static inline bool get_mmap_lock_carefully(struct mm_struct *mm, bool is_user)
|
||||||
|
{
|
||||||
|
if (likely(mmap_read_trylock(mm)))
|
||||||
|
return true;
|
||||||
|
|
||||||
|
if (!is_user)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return !mmap_read_lock_killable(mm);
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline bool mmap_upgrade_trylock(struct mm_struct *mm)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* We don't have this operation yet.
|
||||||
|
*
|
||||||
|
* It should be easy enough to do: it's basically a
|
||||||
|
* atomic_long_try_cmpxchg_acquire()
|
||||||
|
* from RWSEM_READER_BIAS -> RWSEM_WRITER_LOCKED, but
|
||||||
|
* it also needs the proper lockdep magic etc.
|
||||||
|
*/
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline bool upgrade_mmap_lock_carefully(struct mm_struct *mm, bool is_user)
|
||||||
|
{
|
||||||
|
mmap_read_unlock(mm);
|
||||||
|
if (!is_user)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return !mmap_write_lock_killable(mm);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Helper for page fault handling.
|
||||||
|
*
|
||||||
|
* This is kind of equivalend to "mmap_read_lock()" followed
|
||||||
|
* by "find_extend_vma()", except it's a lot more careful about
|
||||||
|
* the locking (and will drop the lock on failure).
|
||||||
|
*
|
||||||
|
* For example, if we have a kernel bug that causes a page
|
||||||
|
* fault, we don't want to just use mmap_read_lock() to get
|
||||||
|
* the mm lock, because that would deadlock if the bug were
|
||||||
|
* to happen while we're holding the mm lock for writing.
|
||||||
|
*
|
||||||
|
* So this checks the exception tables on kernel faults in
|
||||||
|
* order to only do this all for instructions that are actually
|
||||||
|
* expected to fault.
|
||||||
|
*
|
||||||
|
* We can also actually take the mm lock for writing if we
|
||||||
|
* need to extend the vma, which helps the VM layer a lot.
|
||||||
|
*/
|
||||||
|
static struct vm_area_struct *
|
||||||
|
um_lock_mm_and_find_vma(struct mm_struct *mm,
|
||||||
|
unsigned long addr, bool is_user)
|
||||||
|
{
|
||||||
|
struct vm_area_struct *vma;
|
||||||
|
|
||||||
|
if (!get_mmap_lock_carefully(mm, is_user))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
vma = find_vma(mm, addr);
|
||||||
|
if (likely(vma && (vma->vm_start <= addr)))
|
||||||
|
return vma;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Well, dang. We might still be successful, but only
|
||||||
|
* if we can extend a vma to do so.
|
||||||
|
*/
|
||||||
|
if (!vma || !(vma->vm_flags & VM_GROWSDOWN)) {
|
||||||
|
mmap_read_unlock(mm);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We can try to upgrade the mmap lock atomically,
|
||||||
|
* in which case we can continue to use the vma
|
||||||
|
* we already looked up.
|
||||||
|
*
|
||||||
|
* Otherwise we'll have to drop the mmap lock and
|
||||||
|
* re-take it, and also look up the vma again,
|
||||||
|
* re-checking it.
|
||||||
|
*/
|
||||||
|
if (!mmap_upgrade_trylock(mm)) {
|
||||||
|
if (!upgrade_mmap_lock_carefully(mm, is_user))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
vma = find_vma(mm, addr);
|
||||||
|
if (!vma)
|
||||||
|
goto fail;
|
||||||
|
if (vma->vm_start <= addr)
|
||||||
|
goto success;
|
||||||
|
if (!(vma->vm_flags & VM_GROWSDOWN))
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (expand_stack_locked(vma, addr))
|
||||||
|
goto fail;
|
||||||
|
|
||||||
|
success:
|
||||||
|
mmap_write_downgrade(mm);
|
||||||
|
return vma;
|
||||||
|
|
||||||
|
fail:
|
||||||
|
mmap_write_unlock(mm);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Note this is constrained to return 0, -EFAULT, -EACCES, -ENOMEM by
|
* Note this is constrained to return 0, -EFAULT, -EACCES, -ENOMEM by
|
||||||
* segv().
|
* segv().
|
||||||
@@ -43,21 +159,10 @@ int handle_page_fault(unsigned long address, unsigned long ip,
|
|||||||
if (is_user)
|
if (is_user)
|
||||||
flags |= FAULT_FLAG_USER;
|
flags |= FAULT_FLAG_USER;
|
||||||
retry:
|
retry:
|
||||||
mmap_read_lock(mm);
|
vma = um_lock_mm_and_find_vma(mm, address, is_user);
|
||||||
vma = find_vma(mm, address);
|
|
||||||
if (!vma)
|
|
||||||
goto out;
|
|
||||||
if (vma->vm_start <= address)
|
|
||||||
goto good_area;
|
|
||||||
if (!(vma->vm_flags & VM_GROWSDOWN))
|
|
||||||
goto out;
|
|
||||||
if (is_user && !ARCH_IS_STACKGROW(address))
|
|
||||||
goto out;
|
|
||||||
vma = expand_stack(mm, address);
|
|
||||||
if (!vma)
|
if (!vma)
|
||||||
goto out_nosemaphore;
|
goto out_nosemaphore;
|
||||||
|
|
||||||
good_area:
|
|
||||||
*code_out = SEGV_ACCERR;
|
*code_out = SEGV_ACCERR;
|
||||||
if (is_write) {
|
if (is_write) {
|
||||||
if (!(vma->vm_flags & VM_WRITE))
|
if (!(vma->vm_flags & VM_WRITE))
|
||||||
|
@@ -10,8 +10,7 @@
|
|||||||
#include <assert.h>
|
#include <assert.h>
|
||||||
#include <unistd.h>
|
#include <unistd.h>
|
||||||
#include <stdarg.h>
|
#include <stdarg.h>
|
||||||
|
#include <linux/kallsyms.h>
|
||||||
#define unlikely(cond) (cond)
|
|
||||||
|
|
||||||
#include <asm/insn.h>
|
#include <asm/insn.h>
|
||||||
#include <inat.c>
|
#include <inat.c>
|
||||||
@@ -106,7 +105,7 @@ static void parse_args(int argc, char **argv)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#define BUFSIZE 256
|
#define BUFSIZE (256 + KSYM_NAME_LEN)
|
||||||
|
|
||||||
int main(int argc, char **argv)
|
int main(int argc, char **argv)
|
||||||
{
|
{
|
||||||
|
@@ -20,6 +20,9 @@
|
|||||||
*/
|
*/
|
||||||
extern __wsum csum_partial(const void *buff, int len, __wsum sum);
|
extern __wsum csum_partial(const void *buff, int len, __wsum sum);
|
||||||
|
|
||||||
|
/* Do not call this directly. Declared for export type visibility. */
|
||||||
|
extern __visible __wsum csum_partial_copy_generic(const void *src, void *dst, int len);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* csum_fold - Fold and invert a 32bit checksum.
|
* csum_fold - Fold and invert a 32bit checksum.
|
||||||
* sum: 32bit unfolded sum
|
* sum: 32bit unfolded sum
|
||||||
|
@@ -1653,6 +1653,13 @@ static int find_pos_and_ways(struct cxl_port *port, struct range *range,
|
|||||||
}
|
}
|
||||||
put_device(dev);
|
put_device(dev);
|
||||||
|
|
||||||
|
if (rc)
|
||||||
|
dev_err(port->uport_dev,
|
||||||
|
"failed to find %s:%s in target list of %s\n",
|
||||||
|
dev_name(&port->dev),
|
||||||
|
dev_name(port->parent_dport->dport_dev),
|
||||||
|
dev_name(&cxlsd->cxld.dev));
|
||||||
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -354,7 +354,9 @@ static void idxd_cdev_evl_drain_pasid(struct idxd_wq *wq, u32 pasid)
|
|||||||
set_bit(h, evl->bmap);
|
set_bit(h, evl->bmap);
|
||||||
h = (h + 1) % size;
|
h = (h + 1) % size;
|
||||||
}
|
}
|
||||||
|
if (wq->wq)
|
||||||
drain_workqueue(wq->wq);
|
drain_workqueue(wq->wq);
|
||||||
|
|
||||||
mutex_unlock(&evl->lock);
|
mutex_unlock(&evl->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -2900,6 +2900,8 @@ static int xilinx_dma_chan_probe(struct xilinx_dma_device *xdev,
|
|||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
xdev->common.directions |= chan->direction;
|
||||||
|
|
||||||
/* Request the interrupt */
|
/* Request the interrupt */
|
||||||
chan->irq = of_irq_get(node, chan->tdest);
|
chan->irq = of_irq_get(node, chan->tdest);
|
||||||
if (chan->irq < 0)
|
if (chan->irq < 0)
|
||||||
|
@@ -1475,7 +1475,9 @@ static int umc_get_cs_mode(int dimm, u8 ctrl, struct amd64_pvt *pvt)
|
|||||||
if (csrow_enabled(2 * dimm + 1, ctrl, pvt))
|
if (csrow_enabled(2 * dimm + 1, ctrl, pvt))
|
||||||
cs_mode |= CS_ODD_PRIMARY;
|
cs_mode |= CS_ODD_PRIMARY;
|
||||||
|
|
||||||
/* Asymmetric dual-rank DIMM support. */
|
if (csrow_sec_enabled(2 * dimm, ctrl, pvt))
|
||||||
|
cs_mode |= CS_EVEN_SECONDARY;
|
||||||
|
|
||||||
if (csrow_sec_enabled(2 * dimm + 1, ctrl, pvt))
|
if (csrow_sec_enabled(2 * dimm + 1, ctrl, pvt))
|
||||||
cs_mode |= CS_ODD_SECONDARY;
|
cs_mode |= CS_ODD_SECONDARY;
|
||||||
|
|
||||||
@@ -1496,12 +1498,13 @@ static int umc_get_cs_mode(int dimm, u8 ctrl, struct amd64_pvt *pvt)
|
|||||||
return cs_mode;
|
return cs_mode;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __addr_mask_to_cs_size(u32 addr_mask_orig, unsigned int cs_mode,
|
static int calculate_cs_size(u32 mask, unsigned int cs_mode)
|
||||||
int csrow_nr, int dimm)
|
|
||||||
{
|
{
|
||||||
u32 msb, weight, num_zero_bits;
|
int msb, weight, num_zero_bits;
|
||||||
u32 addr_mask_deinterleaved;
|
u32 deinterleaved_mask;
|
||||||
int size = 0;
|
|
||||||
|
if (!mask)
|
||||||
|
return 0;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The number of zero bits in the mask is equal to the number of bits
|
* The number of zero bits in the mask is equal to the number of bits
|
||||||
@@ -1514,19 +1517,30 @@ static int __addr_mask_to_cs_size(u32 addr_mask_orig, unsigned int cs_mode,
|
|||||||
* without swapping with the most significant bit. This can be handled
|
* without swapping with the most significant bit. This can be handled
|
||||||
* by keeping the MSB where it is and ignoring the single zero bit.
|
* by keeping the MSB where it is and ignoring the single zero bit.
|
||||||
*/
|
*/
|
||||||
msb = fls(addr_mask_orig) - 1;
|
msb = fls(mask) - 1;
|
||||||
weight = hweight_long(addr_mask_orig);
|
weight = hweight_long(mask);
|
||||||
num_zero_bits = msb - weight - !!(cs_mode & CS_3R_INTERLEAVE);
|
num_zero_bits = msb - weight - !!(cs_mode & CS_3R_INTERLEAVE);
|
||||||
|
|
||||||
/* Take the number of zero bits off from the top of the mask. */
|
/* Take the number of zero bits off from the top of the mask. */
|
||||||
addr_mask_deinterleaved = GENMASK_ULL(msb - num_zero_bits, 1);
|
deinterleaved_mask = GENMASK(msb - num_zero_bits, 1);
|
||||||
|
edac_dbg(1, " Deinterleaved AddrMask: 0x%x\n", deinterleaved_mask);
|
||||||
|
|
||||||
|
return (deinterleaved_mask >> 2) + 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int __addr_mask_to_cs_size(u32 addr_mask, u32 addr_mask_sec,
|
||||||
|
unsigned int cs_mode, int csrow_nr, int dimm)
|
||||||
|
{
|
||||||
|
int size;
|
||||||
|
|
||||||
edac_dbg(1, "CS%d DIMM%d AddrMasks:\n", csrow_nr, dimm);
|
edac_dbg(1, "CS%d DIMM%d AddrMasks:\n", csrow_nr, dimm);
|
||||||
edac_dbg(1, " Original AddrMask: 0x%x\n", addr_mask_orig);
|
edac_dbg(1, " Primary AddrMask: 0x%x\n", addr_mask);
|
||||||
edac_dbg(1, " Deinterleaved AddrMask: 0x%x\n", addr_mask_deinterleaved);
|
|
||||||
|
|
||||||
/* Register [31:1] = Address [39:9]. Size is in kBs here. */
|
/* Register [31:1] = Address [39:9]. Size is in kBs here. */
|
||||||
size = (addr_mask_deinterleaved >> 2) + 1;
|
size = calculate_cs_size(addr_mask, cs_mode);
|
||||||
|
|
||||||
|
edac_dbg(1, " Secondary AddrMask: 0x%x\n", addr_mask_sec);
|
||||||
|
size += calculate_cs_size(addr_mask_sec, cs_mode);
|
||||||
|
|
||||||
/* Return size in MBs. */
|
/* Return size in MBs. */
|
||||||
return size >> 10;
|
return size >> 10;
|
||||||
@@ -1535,8 +1549,8 @@ static int __addr_mask_to_cs_size(u32 addr_mask_orig, unsigned int cs_mode,
|
|||||||
static int umc_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc,
|
static int umc_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc,
|
||||||
unsigned int cs_mode, int csrow_nr)
|
unsigned int cs_mode, int csrow_nr)
|
||||||
{
|
{
|
||||||
|
u32 addr_mask = 0, addr_mask_sec = 0;
|
||||||
int cs_mask_nr = csrow_nr;
|
int cs_mask_nr = csrow_nr;
|
||||||
u32 addr_mask_orig;
|
|
||||||
int dimm, size = 0;
|
int dimm, size = 0;
|
||||||
|
|
||||||
/* No Chip Selects are enabled. */
|
/* No Chip Selects are enabled. */
|
||||||
@@ -1574,13 +1588,13 @@ static int umc_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc,
|
|||||||
if (!pvt->flags.zn_regs_v2)
|
if (!pvt->flags.zn_regs_v2)
|
||||||
cs_mask_nr >>= 1;
|
cs_mask_nr >>= 1;
|
||||||
|
|
||||||
/* Asymmetric dual-rank DIMM support. */
|
if (cs_mode & (CS_EVEN_PRIMARY | CS_ODD_PRIMARY))
|
||||||
if ((csrow_nr & 1) && (cs_mode & CS_ODD_SECONDARY))
|
addr_mask = pvt->csels[umc].csmasks[cs_mask_nr];
|
||||||
addr_mask_orig = pvt->csels[umc].csmasks_sec[cs_mask_nr];
|
|
||||||
else
|
|
||||||
addr_mask_orig = pvt->csels[umc].csmasks[cs_mask_nr];
|
|
||||||
|
|
||||||
return __addr_mask_to_cs_size(addr_mask_orig, cs_mode, csrow_nr, dimm);
|
if (cs_mode & (CS_EVEN_SECONDARY | CS_ODD_SECONDARY))
|
||||||
|
addr_mask_sec = pvt->csels[umc].csmasks_sec[cs_mask_nr];
|
||||||
|
|
||||||
|
return __addr_mask_to_cs_size(addr_mask, addr_mask_sec, cs_mode, csrow_nr, dimm);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void umc_debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
|
static void umc_debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
|
||||||
@@ -3773,9 +3787,10 @@ static void gpu_get_err_info(struct mce *m, struct err_info *err)
|
|||||||
static int gpu_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc,
|
static int gpu_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc,
|
||||||
unsigned int cs_mode, int csrow_nr)
|
unsigned int cs_mode, int csrow_nr)
|
||||||
{
|
{
|
||||||
u32 addr_mask_orig = pvt->csels[umc].csmasks[csrow_nr];
|
u32 addr_mask = pvt->csels[umc].csmasks[csrow_nr];
|
||||||
|
u32 addr_mask_sec = pvt->csels[umc].csmasks_sec[csrow_nr];
|
||||||
|
|
||||||
return __addr_mask_to_cs_size(addr_mask_orig, cs_mode, csrow_nr, csrow_nr >> 1);
|
return __addr_mask_to_cs_size(addr_mask, addr_mask_sec, cs_mode, csrow_nr, csrow_nr >> 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void gpu_debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
|
static void gpu_debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
|
||||||
|
@@ -1571,6 +1571,39 @@ static int scmi_common_get_max_msg_size(const struct scmi_protocol_handle *ph)
|
|||||||
return info->desc->max_msg_size;
|
return info->desc->max_msg_size;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* scmi_protocol_msg_check - Check protocol message attributes
|
||||||
|
*
|
||||||
|
* @ph: A reference to the protocol handle.
|
||||||
|
* @message_id: The ID of the message to check.
|
||||||
|
* @attributes: A parameter to optionally return the retrieved message
|
||||||
|
* attributes, in case of Success.
|
||||||
|
*
|
||||||
|
* An helper to check protocol message attributes for a specific protocol
|
||||||
|
* and message pair.
|
||||||
|
*
|
||||||
|
* Return: 0 on SUCCESS
|
||||||
|
*/
|
||||||
|
static int scmi_protocol_msg_check(const struct scmi_protocol_handle *ph,
|
||||||
|
u32 message_id, u32 *attributes)
|
||||||
|
{
|
||||||
|
int ret;
|
||||||
|
struct scmi_xfer *t;
|
||||||
|
|
||||||
|
ret = xfer_get_init(ph, PROTOCOL_MESSAGE_ATTRIBUTES,
|
||||||
|
sizeof(__le32), 0, &t);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
|
put_unaligned_le32(message_id, t->tx.buf);
|
||||||
|
ret = do_xfer(ph, t);
|
||||||
|
if (!ret && attributes)
|
||||||
|
*attributes = get_unaligned_le32(t->rx.buf);
|
||||||
|
xfer_put(ph, t);
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* struct scmi_iterator - Iterator descriptor
|
* struct scmi_iterator - Iterator descriptor
|
||||||
* @msg: A reference to the message TX buffer; filled by @prepare_message with
|
* @msg: A reference to the message TX buffer; filled by @prepare_message with
|
||||||
@@ -1712,6 +1745,7 @@ scmi_common_fastchannel_init(const struct scmi_protocol_handle *ph,
|
|||||||
int ret;
|
int ret;
|
||||||
u32 flags;
|
u32 flags;
|
||||||
u64 phys_addr;
|
u64 phys_addr;
|
||||||
|
u32 attributes;
|
||||||
u8 size;
|
u8 size;
|
||||||
void __iomem *addr;
|
void __iomem *addr;
|
||||||
struct scmi_xfer *t;
|
struct scmi_xfer *t;
|
||||||
@@ -1720,6 +1754,15 @@ scmi_common_fastchannel_init(const struct scmi_protocol_handle *ph,
|
|||||||
struct scmi_msg_resp_desc_fc *resp;
|
struct scmi_msg_resp_desc_fc *resp;
|
||||||
const struct scmi_protocol_instance *pi = ph_to_pi(ph);
|
const struct scmi_protocol_instance *pi = ph_to_pi(ph);
|
||||||
|
|
||||||
|
/* Check if the MSG_ID supports fastchannel */
|
||||||
|
ret = scmi_protocol_msg_check(ph, message_id, &attributes);
|
||||||
|
if (ret || !MSG_SUPPORTS_FASTCHANNEL(attributes)) {
|
||||||
|
dev_dbg(ph->dev,
|
||||||
|
"Skip FC init for 0x%02X/%d domain:%d - ret:%d\n",
|
||||||
|
pi->proto->id, message_id, domain, ret);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
if (!p_addr) {
|
if (!p_addr) {
|
||||||
ret = -EINVAL;
|
ret = -EINVAL;
|
||||||
goto err_out;
|
goto err_out;
|
||||||
@@ -1849,6 +1892,7 @@ static const struct scmi_proto_helpers_ops helpers_ops = {
|
|||||||
.get_max_msg_size = scmi_common_get_max_msg_size,
|
.get_max_msg_size = scmi_common_get_max_msg_size,
|
||||||
.iter_response_init = scmi_iterator_init,
|
.iter_response_init = scmi_iterator_init,
|
||||||
.iter_response_run = scmi_iterator_run,
|
.iter_response_run = scmi_iterator_run,
|
||||||
|
.protocol_msg_check = scmi_protocol_msg_check,
|
||||||
.fastchannel_init = scmi_common_fastchannel_init,
|
.fastchannel_init = scmi_common_fastchannel_init,
|
||||||
.fastchannel_db_ring = scmi_common_fastchannel_db_ring,
|
.fastchannel_db_ring = scmi_common_fastchannel_db_ring,
|
||||||
};
|
};
|
||||||
|
@@ -29,6 +29,8 @@
|
|||||||
#define PROTOCOL_REV_MAJOR(x) ((u16)(FIELD_GET(PROTOCOL_REV_MAJOR_MASK, (x))))
|
#define PROTOCOL_REV_MAJOR(x) ((u16)(FIELD_GET(PROTOCOL_REV_MAJOR_MASK, (x))))
|
||||||
#define PROTOCOL_REV_MINOR(x) ((u16)(FIELD_GET(PROTOCOL_REV_MINOR_MASK, (x))))
|
#define PROTOCOL_REV_MINOR(x) ((u16)(FIELD_GET(PROTOCOL_REV_MINOR_MASK, (x))))
|
||||||
|
|
||||||
|
#define MSG_SUPPORTS_FASTCHANNEL(x) ((x) & BIT(0))
|
||||||
|
|
||||||
enum scmi_common_cmd {
|
enum scmi_common_cmd {
|
||||||
PROTOCOL_VERSION = 0x0,
|
PROTOCOL_VERSION = 0x0,
|
||||||
PROTOCOL_ATTRIBUTES = 0x1,
|
PROTOCOL_ATTRIBUTES = 0x1,
|
||||||
@@ -251,6 +253,8 @@ struct scmi_fc_info {
|
|||||||
* provided in @ops.
|
* provided in @ops.
|
||||||
* @iter_response_run: A common helper to trigger the run of a previously
|
* @iter_response_run: A common helper to trigger the run of a previously
|
||||||
* initialized iterator.
|
* initialized iterator.
|
||||||
|
* @protocol_msg_check: A common helper to check is a specific protocol message
|
||||||
|
* is supported.
|
||||||
* @fastchannel_init: A common helper used to initialize FC descriptors by
|
* @fastchannel_init: A common helper used to initialize FC descriptors by
|
||||||
* gathering FC descriptions from the SCMI platform server.
|
* gathering FC descriptions from the SCMI platform server.
|
||||||
* @fastchannel_db_ring: A common helper to ring a FC doorbell.
|
* @fastchannel_db_ring: A common helper to ring a FC doorbell.
|
||||||
@@ -265,6 +269,8 @@ struct scmi_proto_helpers_ops {
|
|||||||
unsigned int max_resources, u8 msg_id,
|
unsigned int max_resources, u8 msg_id,
|
||||||
size_t tx_size, void *priv);
|
size_t tx_size, void *priv);
|
||||||
int (*iter_response_run)(void *iter);
|
int (*iter_response_run)(void *iter);
|
||||||
|
int (*protocol_msg_check)(const struct scmi_protocol_handle *ph,
|
||||||
|
u32 message_id, u32 *attributes);
|
||||||
void (*fastchannel_init)(const struct scmi_protocol_handle *ph,
|
void (*fastchannel_init)(const struct scmi_protocol_handle *ph,
|
||||||
u8 describe_id, u32 message_id,
|
u8 describe_id, u32 message_id,
|
||||||
u32 valid_size, u32 domain,
|
u32 valid_size, u32 domain,
|
||||||
|
@@ -1890,7 +1890,7 @@ no_preempt:
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
job = to_amdgpu_job(s_job);
|
job = to_amdgpu_job(s_job);
|
||||||
if (preempted && (&job->hw_fence) == fence)
|
if (preempted && (&job->hw_fence.base) == fence)
|
||||||
/* mark the job as preempted */
|
/* mark the job as preempted */
|
||||||
job->preemption_status |= AMDGPU_IB_PREEMPTED;
|
job->preemption_status |= AMDGPU_IB_PREEMPTED;
|
||||||
}
|
}
|
||||||
|
@@ -5367,7 +5367,7 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev,
|
|||||||
*
|
*
|
||||||
* job->base holds a reference to parent fence
|
* job->base holds a reference to parent fence
|
||||||
*/
|
*/
|
||||||
if (job && dma_fence_is_signaled(&job->hw_fence)) {
|
if (job && dma_fence_is_signaled(&job->hw_fence.base)) {
|
||||||
job_signaled = true;
|
job_signaled = true;
|
||||||
dev_info(adev->dev, "Guilty job already signaled, skipping HW reset");
|
dev_info(adev->dev, "Guilty job already signaled, skipping HW reset");
|
||||||
goto skip_hw_reset;
|
goto skip_hw_reset;
|
||||||
|
@@ -41,22 +41,6 @@
|
|||||||
#include "amdgpu_trace.h"
|
#include "amdgpu_trace.h"
|
||||||
#include "amdgpu_reset.h"
|
#include "amdgpu_reset.h"
|
||||||
|
|
||||||
/*
|
|
||||||
* Fences mark an event in the GPUs pipeline and are used
|
|
||||||
* for GPU/CPU synchronization. When the fence is written,
|
|
||||||
* it is expected that all buffers associated with that fence
|
|
||||||
* are no longer in use by the associated ring on the GPU and
|
|
||||||
* that the relevant GPU caches have been flushed.
|
|
||||||
*/
|
|
||||||
|
|
||||||
struct amdgpu_fence {
|
|
||||||
struct dma_fence base;
|
|
||||||
|
|
||||||
/* RB, DMA, etc. */
|
|
||||||
struct amdgpu_ring *ring;
|
|
||||||
ktime_t start_timestamp;
|
|
||||||
};
|
|
||||||
|
|
||||||
static struct kmem_cache *amdgpu_fence_slab;
|
static struct kmem_cache *amdgpu_fence_slab;
|
||||||
|
|
||||||
int amdgpu_fence_slab_init(void)
|
int amdgpu_fence_slab_init(void)
|
||||||
@@ -153,12 +137,12 @@ int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amd
|
|||||||
am_fence = kmem_cache_alloc(amdgpu_fence_slab, GFP_ATOMIC);
|
am_fence = kmem_cache_alloc(amdgpu_fence_slab, GFP_ATOMIC);
|
||||||
if (am_fence == NULL)
|
if (am_fence == NULL)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
fence = &am_fence->base;
|
|
||||||
am_fence->ring = ring;
|
|
||||||
} else {
|
} else {
|
||||||
/* take use of job-embedded fence */
|
/* take use of job-embedded fence */
|
||||||
fence = &job->hw_fence;
|
am_fence = &job->hw_fence;
|
||||||
}
|
}
|
||||||
|
fence = &am_fence->base;
|
||||||
|
am_fence->ring = ring;
|
||||||
|
|
||||||
seq = ++ring->fence_drv.sync_seq;
|
seq = ++ring->fence_drv.sync_seq;
|
||||||
if (job && job->job_run_counter) {
|
if (job && job->job_run_counter) {
|
||||||
@@ -719,7 +703,7 @@ void amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring *ring)
|
|||||||
* it right here or we won't be able to track them in fence_drv
|
* it right here or we won't be able to track them in fence_drv
|
||||||
* and they will remain unsignaled during sa_bo free.
|
* and they will remain unsignaled during sa_bo free.
|
||||||
*/
|
*/
|
||||||
job = container_of(old, struct amdgpu_job, hw_fence);
|
job = container_of(old, struct amdgpu_job, hw_fence.base);
|
||||||
if (!job->base.s_fence && !dma_fence_is_signaled(old))
|
if (!job->base.s_fence && !dma_fence_is_signaled(old))
|
||||||
dma_fence_signal(old);
|
dma_fence_signal(old);
|
||||||
RCU_INIT_POINTER(*ptr, NULL);
|
RCU_INIT_POINTER(*ptr, NULL);
|
||||||
@@ -781,7 +765,7 @@ static const char *amdgpu_fence_get_timeline_name(struct dma_fence *f)
|
|||||||
|
|
||||||
static const char *amdgpu_job_fence_get_timeline_name(struct dma_fence *f)
|
static const char *amdgpu_job_fence_get_timeline_name(struct dma_fence *f)
|
||||||
{
|
{
|
||||||
struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence);
|
struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence.base);
|
||||||
|
|
||||||
return (const char *)to_amdgpu_ring(job->base.sched)->name;
|
return (const char *)to_amdgpu_ring(job->base.sched)->name;
|
||||||
}
|
}
|
||||||
@@ -811,7 +795,7 @@ static bool amdgpu_fence_enable_signaling(struct dma_fence *f)
|
|||||||
*/
|
*/
|
||||||
static bool amdgpu_job_fence_enable_signaling(struct dma_fence *f)
|
static bool amdgpu_job_fence_enable_signaling(struct dma_fence *f)
|
||||||
{
|
{
|
||||||
struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence);
|
struct amdgpu_job *job = container_of(f, struct amdgpu_job, hw_fence.base);
|
||||||
|
|
||||||
if (!timer_pending(&to_amdgpu_ring(job->base.sched)->fence_drv.fallback_timer))
|
if (!timer_pending(&to_amdgpu_ring(job->base.sched)->fence_drv.fallback_timer))
|
||||||
amdgpu_fence_schedule_fallback(to_amdgpu_ring(job->base.sched));
|
amdgpu_fence_schedule_fallback(to_amdgpu_ring(job->base.sched));
|
||||||
@@ -846,7 +830,7 @@ static void amdgpu_job_fence_free(struct rcu_head *rcu)
|
|||||||
struct dma_fence *f = container_of(rcu, struct dma_fence, rcu);
|
struct dma_fence *f = container_of(rcu, struct dma_fence, rcu);
|
||||||
|
|
||||||
/* free job if fence has a parent job */
|
/* free job if fence has a parent job */
|
||||||
kfree(container_of(f, struct amdgpu_job, hw_fence));
|
kfree(container_of(f, struct amdgpu_job, hw_fence.base));
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@@ -165,8 +165,8 @@ void amdgpu_job_free_resources(struct amdgpu_job *job)
|
|||||||
/* Check if any fences where initialized */
|
/* Check if any fences where initialized */
|
||||||
if (job->base.s_fence && job->base.s_fence->finished.ops)
|
if (job->base.s_fence && job->base.s_fence->finished.ops)
|
||||||
f = &job->base.s_fence->finished;
|
f = &job->base.s_fence->finished;
|
||||||
else if (job->hw_fence.ops)
|
else if (job->hw_fence.base.ops)
|
||||||
f = &job->hw_fence;
|
f = &job->hw_fence.base;
|
||||||
else
|
else
|
||||||
f = NULL;
|
f = NULL;
|
||||||
|
|
||||||
@@ -183,10 +183,10 @@ static void amdgpu_job_free_cb(struct drm_sched_job *s_job)
|
|||||||
amdgpu_sync_free(&job->explicit_sync);
|
amdgpu_sync_free(&job->explicit_sync);
|
||||||
|
|
||||||
/* only put the hw fence if has embedded fence */
|
/* only put the hw fence if has embedded fence */
|
||||||
if (!job->hw_fence.ops)
|
if (!job->hw_fence.base.ops)
|
||||||
kfree(job);
|
kfree(job);
|
||||||
else
|
else
|
||||||
dma_fence_put(&job->hw_fence);
|
dma_fence_put(&job->hw_fence.base);
|
||||||
}
|
}
|
||||||
|
|
||||||
void amdgpu_job_set_gang_leader(struct amdgpu_job *job,
|
void amdgpu_job_set_gang_leader(struct amdgpu_job *job,
|
||||||
@@ -215,10 +215,10 @@ void amdgpu_job_free(struct amdgpu_job *job)
|
|||||||
if (job->gang_submit != &job->base.s_fence->scheduled)
|
if (job->gang_submit != &job->base.s_fence->scheduled)
|
||||||
dma_fence_put(job->gang_submit);
|
dma_fence_put(job->gang_submit);
|
||||||
|
|
||||||
if (!job->hw_fence.ops)
|
if (!job->hw_fence.base.ops)
|
||||||
kfree(job);
|
kfree(job);
|
||||||
else
|
else
|
||||||
dma_fence_put(&job->hw_fence);
|
dma_fence_put(&job->hw_fence.base);
|
||||||
}
|
}
|
||||||
|
|
||||||
struct dma_fence *amdgpu_job_submit(struct amdgpu_job *job)
|
struct dma_fence *amdgpu_job_submit(struct amdgpu_job *job)
|
||||||
|
@@ -48,7 +48,7 @@ struct amdgpu_job {
|
|||||||
struct drm_sched_job base;
|
struct drm_sched_job base;
|
||||||
struct amdgpu_vm *vm;
|
struct amdgpu_vm *vm;
|
||||||
struct amdgpu_sync explicit_sync;
|
struct amdgpu_sync explicit_sync;
|
||||||
struct dma_fence hw_fence;
|
struct amdgpu_fence hw_fence;
|
||||||
struct dma_fence *gang_submit;
|
struct dma_fence *gang_submit;
|
||||||
uint32_t preamble_status;
|
uint32_t preamble_status;
|
||||||
uint32_t preemption_status;
|
uint32_t preemption_status;
|
||||||
|
@@ -123,6 +123,22 @@ struct amdgpu_fence_driver {
|
|||||||
struct dma_fence **fences;
|
struct dma_fence **fences;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Fences mark an event in the GPUs pipeline and are used
|
||||||
|
* for GPU/CPU synchronization. When the fence is written,
|
||||||
|
* it is expected that all buffers associated with that fence
|
||||||
|
* are no longer in use by the associated ring on the GPU and
|
||||||
|
* that the relevant GPU caches have been flushed.
|
||||||
|
*/
|
||||||
|
|
||||||
|
struct amdgpu_fence {
|
||||||
|
struct dma_fence base;
|
||||||
|
|
||||||
|
/* RB, DMA, etc. */
|
||||||
|
struct amdgpu_ring *ring;
|
||||||
|
ktime_t start_timestamp;
|
||||||
|
};
|
||||||
|
|
||||||
extern const struct drm_sched_backend_ops amdgpu_sched_ops;
|
extern const struct drm_sched_backend_ops amdgpu_sched_ops;
|
||||||
|
|
||||||
void amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring *ring);
|
void amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring *ring);
|
||||||
|
@@ -28,6 +28,10 @@
|
|||||||
#include "amdgpu.h"
|
#include "amdgpu.h"
|
||||||
#include "amdgpu_ucode.h"
|
#include "amdgpu_ucode.h"
|
||||||
|
|
||||||
|
static const struct kicker_device kicker_device_list[] = {
|
||||||
|
{0x744B, 0x00},
|
||||||
|
};
|
||||||
|
|
||||||
static void amdgpu_ucode_print_common_hdr(const struct common_firmware_header *hdr)
|
static void amdgpu_ucode_print_common_hdr(const struct common_firmware_header *hdr)
|
||||||
{
|
{
|
||||||
DRM_DEBUG("size_bytes: %u\n", le32_to_cpu(hdr->size_bytes));
|
DRM_DEBUG("size_bytes: %u\n", le32_to_cpu(hdr->size_bytes));
|
||||||
@@ -1268,6 +1272,19 @@ static const char *amdgpu_ucode_legacy_naming(struct amdgpu_device *adev, int bl
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool amdgpu_is_kicker_fw(struct amdgpu_device *adev)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
|
for (i = 0; i < ARRAY_SIZE(kicker_device_list); i++) {
|
||||||
|
if (adev->pdev->device == kicker_device_list[i].device &&
|
||||||
|
adev->pdev->revision == kicker_device_list[i].revision)
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
void amdgpu_ucode_ip_version_decode(struct amdgpu_device *adev, int block_type, char *ucode_prefix, int len)
|
void amdgpu_ucode_ip_version_decode(struct amdgpu_device *adev, int block_type, char *ucode_prefix, int len)
|
||||||
{
|
{
|
||||||
int maj, min, rev;
|
int maj, min, rev;
|
||||||
|
@@ -536,6 +536,11 @@ struct amdgpu_firmware {
|
|||||||
uint64_t fw_buf_mc;
|
uint64_t fw_buf_mc;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
struct kicker_device{
|
||||||
|
unsigned short device;
|
||||||
|
u8 revision;
|
||||||
|
};
|
||||||
|
|
||||||
void amdgpu_ucode_print_mc_hdr(const struct common_firmware_header *hdr);
|
void amdgpu_ucode_print_mc_hdr(const struct common_firmware_header *hdr);
|
||||||
void amdgpu_ucode_print_smc_hdr(const struct common_firmware_header *hdr);
|
void amdgpu_ucode_print_smc_hdr(const struct common_firmware_header *hdr);
|
||||||
void amdgpu_ucode_print_imu_hdr(const struct common_firmware_header *hdr);
|
void amdgpu_ucode_print_imu_hdr(const struct common_firmware_header *hdr);
|
||||||
@@ -562,5 +567,6 @@ amdgpu_ucode_get_load_type(struct amdgpu_device *adev, int load_type);
|
|||||||
const char *amdgpu_ucode_name(enum AMDGPU_UCODE_ID ucode_id);
|
const char *amdgpu_ucode_name(enum AMDGPU_UCODE_ID ucode_id);
|
||||||
|
|
||||||
void amdgpu_ucode_ip_version_decode(struct amdgpu_device *adev, int block_type, char *ucode_prefix, int len);
|
void amdgpu_ucode_ip_version_decode(struct amdgpu_device *adev, int block_type, char *ucode_prefix, int len);
|
||||||
|
bool amdgpu_is_kicker_fw(struct amdgpu_device *adev);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@@ -435,7 +435,7 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
|
|||||||
int r;
|
int r;
|
||||||
|
|
||||||
lpfn = (u64)place->lpfn << PAGE_SHIFT;
|
lpfn = (u64)place->lpfn << PAGE_SHIFT;
|
||||||
if (!lpfn)
|
if (!lpfn || lpfn > man->size)
|
||||||
lpfn = man->size;
|
lpfn = man->size;
|
||||||
|
|
||||||
fpfn = (u64)place->fpfn << PAGE_SHIFT;
|
fpfn = (u64)place->fpfn << PAGE_SHIFT;
|
||||||
|
@@ -1287,6 +1287,7 @@ void kfd_signal_poison_consumed_event(struct kfd_node *dev, u32 pasid)
|
|||||||
user_gpu_id = kfd_process_get_user_gpu_id(p, dev->id);
|
user_gpu_id = kfd_process_get_user_gpu_id(p, dev->id);
|
||||||
if (unlikely(user_gpu_id == -EINVAL)) {
|
if (unlikely(user_gpu_id == -EINVAL)) {
|
||||||
WARN_ONCE(1, "Could not get user_gpu_id from dev->id:%x\n", dev->id);
|
WARN_ONCE(1, "Could not get user_gpu_id from dev->id:%x\n", dev->id);
|
||||||
|
kfd_unref_process(p);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -225,7 +225,7 @@ static int pm_map_queues_v9(struct packet_manager *pm, uint32_t *buffer,
|
|||||||
|
|
||||||
packet->bitfields2.engine_sel =
|
packet->bitfields2.engine_sel =
|
||||||
engine_sel__mes_map_queues__compute_vi;
|
engine_sel__mes_map_queues__compute_vi;
|
||||||
packet->bitfields2.gws_control_queue = q->gws ? 1 : 0;
|
packet->bitfields2.gws_control_queue = q->properties.is_gws ? 1 : 0;
|
||||||
packet->bitfields2.extended_engine_sel =
|
packet->bitfields2.extended_engine_sel =
|
||||||
extended_engine_sel__mes_map_queues__legacy_engine_sel;
|
extended_engine_sel__mes_map_queues__legacy_engine_sel;
|
||||||
packet->bitfields2.queue_type =
|
packet->bitfields2.queue_type =
|
||||||
|
@@ -368,6 +368,9 @@ enum mod_hdcp_status mod_hdcp_hdcp1_enable_encryption(struct mod_hdcp *hdcp)
|
|||||||
struct mod_hdcp_display *display = get_first_active_display(hdcp);
|
struct mod_hdcp_display *display = get_first_active_display(hdcp);
|
||||||
enum mod_hdcp_status status = MOD_HDCP_STATUS_SUCCESS;
|
enum mod_hdcp_status status = MOD_HDCP_STATUS_SUCCESS;
|
||||||
|
|
||||||
|
if (!display)
|
||||||
|
return MOD_HDCP_STATUS_DISPLAY_NOT_FOUND;
|
||||||
|
|
||||||
mutex_lock(&psp->hdcp_context.mutex);
|
mutex_lock(&psp->hdcp_context.mutex);
|
||||||
hdcp_cmd = (struct ta_hdcp_shared_memory *)psp->hdcp_context.context.mem_context.shared_buf;
|
hdcp_cmd = (struct ta_hdcp_shared_memory *)psp->hdcp_context.context.mem_context.shared_buf;
|
||||||
memset(hdcp_cmd, 0, sizeof(struct ta_hdcp_shared_memory));
|
memset(hdcp_cmd, 0, sizeof(struct ta_hdcp_shared_memory));
|
||||||
|
@@ -1868,9 +1868,9 @@ static void ast_mode_config_helper_atomic_commit_tail(struct drm_atomic_state *s
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* Concurrent operations could possibly trigger a call to
|
* Concurrent operations could possibly trigger a call to
|
||||||
* drm_connector_helper_funcs.get_modes by trying to read the
|
* drm_connector_helper_funcs.get_modes by reading the display
|
||||||
* display modes. Protect access to I/O registers by acquiring
|
* modes. Protect access to registers by acquiring the modeset
|
||||||
* the I/O-register lock. Released in atomic_flush().
|
* lock.
|
||||||
*/
|
*/
|
||||||
mutex_lock(&ast->ioregs_lock);
|
mutex_lock(&ast->ioregs_lock);
|
||||||
drm_atomic_helper_commit_tail_rpm(state);
|
drm_atomic_helper_commit_tail_rpm(state);
|
||||||
|
@@ -568,15 +568,18 @@ static int cdns_dsi_check_conf(struct cdns_dsi *dsi,
|
|||||||
struct phy_configure_opts_mipi_dphy *phy_cfg = &output->phy_opts.mipi_dphy;
|
struct phy_configure_opts_mipi_dphy *phy_cfg = &output->phy_opts.mipi_dphy;
|
||||||
unsigned long dsi_hss_hsa_hse_hbp;
|
unsigned long dsi_hss_hsa_hse_hbp;
|
||||||
unsigned int nlanes = output->dev->lanes;
|
unsigned int nlanes = output->dev->lanes;
|
||||||
|
int mode_clock = (mode_valid_check ? mode->clock : mode->crtc_clock);
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
ret = cdns_dsi_mode2cfg(dsi, mode, dsi_cfg, mode_valid_check);
|
ret = cdns_dsi_mode2cfg(dsi, mode, dsi_cfg, mode_valid_check);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
phy_mipi_dphy_get_default_config(mode->crtc_clock * 1000,
|
ret = phy_mipi_dphy_get_default_config(mode_clock * 1000,
|
||||||
mipi_dsi_pixel_format_to_bpp(output->dev->format),
|
mipi_dsi_pixel_format_to_bpp(output->dev->format),
|
||||||
nlanes, phy_cfg);
|
nlanes, phy_cfg);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
ret = cdns_dsi_adjust_phy_config(dsi, dsi_cfg, phy_cfg, mode, mode_valid_check);
|
ret = cdns_dsi_adjust_phy_config(dsi, dsi_cfg, phy_cfg, mode, mode_valid_check);
|
||||||
if (ret)
|
if (ret)
|
||||||
@@ -680,6 +683,11 @@ static void cdns_dsi_bridge_post_disable(struct drm_bridge *bridge)
|
|||||||
struct cdns_dsi_input *input = bridge_to_cdns_dsi_input(bridge);
|
struct cdns_dsi_input *input = bridge_to_cdns_dsi_input(bridge);
|
||||||
struct cdns_dsi *dsi = input_to_dsi(input);
|
struct cdns_dsi *dsi = input_to_dsi(input);
|
||||||
|
|
||||||
|
dsi->phy_initialized = false;
|
||||||
|
dsi->link_initialized = false;
|
||||||
|
phy_power_off(dsi->dphy);
|
||||||
|
phy_exit(dsi->dphy);
|
||||||
|
|
||||||
pm_runtime_put(dsi->base.dev);
|
pm_runtime_put(dsi->base.dev);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -761,7 +769,7 @@ static void cdns_dsi_bridge_enable(struct drm_bridge *bridge)
|
|||||||
struct phy_configure_opts_mipi_dphy *phy_cfg = &output->phy_opts.mipi_dphy;
|
struct phy_configure_opts_mipi_dphy *phy_cfg = &output->phy_opts.mipi_dphy;
|
||||||
unsigned long tx_byte_period;
|
unsigned long tx_byte_period;
|
||||||
struct cdns_dsi_cfg dsi_cfg;
|
struct cdns_dsi_cfg dsi_cfg;
|
||||||
u32 tmp, reg_wakeup, div;
|
u32 tmp, reg_wakeup, div, status;
|
||||||
int nlanes;
|
int nlanes;
|
||||||
|
|
||||||
if (WARN_ON(pm_runtime_get_sync(dsi->base.dev) < 0))
|
if (WARN_ON(pm_runtime_get_sync(dsi->base.dev) < 0))
|
||||||
@@ -778,6 +786,19 @@ static void cdns_dsi_bridge_enable(struct drm_bridge *bridge)
|
|||||||
cdns_dsi_hs_init(dsi);
|
cdns_dsi_hs_init(dsi);
|
||||||
cdns_dsi_init_link(dsi);
|
cdns_dsi_init_link(dsi);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Now that the DSI Link and DSI Phy are initialized,
|
||||||
|
* wait for the CLK and Data Lanes to be ready.
|
||||||
|
*/
|
||||||
|
tmp = CLK_LANE_RDY;
|
||||||
|
for (int i = 0; i < nlanes; i++)
|
||||||
|
tmp |= DATA_LANE_RDY(i);
|
||||||
|
|
||||||
|
if (readl_poll_timeout(dsi->regs + MCTL_MAIN_STS, status,
|
||||||
|
(tmp == (status & tmp)), 100, 500000))
|
||||||
|
dev_err(dsi->base.dev,
|
||||||
|
"Timed Out: DSI-DPhy Clock and Data Lanes not ready.\n");
|
||||||
|
|
||||||
writel(HBP_LEN(dsi_cfg.hbp) | HSA_LEN(dsi_cfg.hsa),
|
writel(HBP_LEN(dsi_cfg.hbp) | HSA_LEN(dsi_cfg.hsa),
|
||||||
dsi->regs + VID_HSIZE1);
|
dsi->regs + VID_HSIZE1);
|
||||||
writel(HFP_LEN(dsi_cfg.hfp) | HACT_LEN(dsi_cfg.hact),
|
writel(HFP_LEN(dsi_cfg.hfp) | HACT_LEN(dsi_cfg.hact),
|
||||||
@@ -952,7 +973,7 @@ static int cdns_dsi_attach(struct mipi_dsi_host *host,
|
|||||||
bridge = drm_panel_bridge_add_typed(panel,
|
bridge = drm_panel_bridge_add_typed(panel,
|
||||||
DRM_MODE_CONNECTOR_DSI);
|
DRM_MODE_CONNECTOR_DSI);
|
||||||
} else {
|
} else {
|
||||||
bridge = of_drm_find_bridge(dev->dev.of_node);
|
bridge = of_drm_find_bridge(np);
|
||||||
if (!bridge)
|
if (!bridge)
|
||||||
bridge = ERR_PTR(-EINVAL);
|
bridge = ERR_PTR(-EINVAL);
|
||||||
}
|
}
|
||||||
@@ -1152,7 +1173,6 @@ static int __maybe_unused cdns_dsi_suspend(struct device *dev)
|
|||||||
clk_disable_unprepare(dsi->dsi_sys_clk);
|
clk_disable_unprepare(dsi->dsi_sys_clk);
|
||||||
clk_disable_unprepare(dsi->dsi_p_clk);
|
clk_disable_unprepare(dsi->dsi_p_clk);
|
||||||
reset_control_assert(dsi->dsi_p_rst);
|
reset_control_assert(dsi->dsi_p_rst);
|
||||||
dsi->link_initialized = false;
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -331,10 +331,16 @@ static void ti_sn65dsi86_enable_comms(struct ti_sn65dsi86 *pdata)
|
|||||||
* 200 ms. We'll assume that the panel driver will have the hardcoded
|
* 200 ms. We'll assume that the panel driver will have the hardcoded
|
||||||
* delay in its prepare and always disable HPD.
|
* delay in its prepare and always disable HPD.
|
||||||
*
|
*
|
||||||
* If HPD somehow makes sense on some future panel we'll have to
|
* For DisplayPort bridge type, we need HPD. So we use the bridge type
|
||||||
* change this to be conditional on someone specifying that HPD should
|
* to conditionally disable HPD.
|
||||||
* be used.
|
* NOTE: The bridge type is set in ti_sn_bridge_probe() but enable_comms()
|
||||||
|
* can be called before. So for DisplayPort, HPD will be enabled once
|
||||||
|
* bridge type is set. We are using bridge type instead of "no-hpd"
|
||||||
|
* property because it is not used properly in devicetree description
|
||||||
|
* and hence is unreliable.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
if (pdata->bridge.type != DRM_MODE_CONNECTOR_DisplayPort)
|
||||||
regmap_update_bits(pdata->regmap, SN_HPD_DISABLE_REG, HPD_DISABLE,
|
regmap_update_bits(pdata->regmap, SN_HPD_DISABLE_REG, HPD_DISABLE,
|
||||||
HPD_DISABLE);
|
HPD_DISABLE);
|
||||||
|
|
||||||
@@ -424,36 +430,8 @@ static int status_show(struct seq_file *s, void *data)
|
|||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
DEFINE_SHOW_ATTRIBUTE(status);
|
DEFINE_SHOW_ATTRIBUTE(status);
|
||||||
|
|
||||||
static void ti_sn65dsi86_debugfs_remove(void *data)
|
|
||||||
{
|
|
||||||
debugfs_remove_recursive(data);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void ti_sn65dsi86_debugfs_init(struct ti_sn65dsi86 *pdata)
|
|
||||||
{
|
|
||||||
struct device *dev = pdata->dev;
|
|
||||||
struct dentry *debugfs;
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
debugfs = debugfs_create_dir(dev_name(dev), NULL);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* We might get an error back if debugfs wasn't enabled in the kernel
|
|
||||||
* so let's just silently return upon failure.
|
|
||||||
*/
|
|
||||||
if (IS_ERR_OR_NULL(debugfs))
|
|
||||||
return;
|
|
||||||
|
|
||||||
ret = devm_add_action_or_reset(dev, ti_sn65dsi86_debugfs_remove, debugfs);
|
|
||||||
if (ret)
|
|
||||||
return;
|
|
||||||
|
|
||||||
debugfs_create_file("status", 0600, debugfs, pdata, &status_fops);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* -----------------------------------------------------------------------------
|
/* -----------------------------------------------------------------------------
|
||||||
* Auxiliary Devices (*not* AUX)
|
* Auxiliary Devices (*not* AUX)
|
||||||
*/
|
*/
|
||||||
@@ -1201,9 +1179,14 @@ static enum drm_connector_status ti_sn_bridge_detect(struct drm_bridge *bridge)
|
|||||||
struct ti_sn65dsi86 *pdata = bridge_to_ti_sn65dsi86(bridge);
|
struct ti_sn65dsi86 *pdata = bridge_to_ti_sn65dsi86(bridge);
|
||||||
int val = 0;
|
int val = 0;
|
||||||
|
|
||||||
pm_runtime_get_sync(pdata->dev);
|
/*
|
||||||
|
* Runtime reference is grabbed in ti_sn_bridge_hpd_enable()
|
||||||
|
* as the chip won't report HPD just after being powered on.
|
||||||
|
* HPD_DEBOUNCED_STATE reflects correct state only after the
|
||||||
|
* debounce time (~100-400 ms).
|
||||||
|
*/
|
||||||
|
|
||||||
regmap_read(pdata->regmap, SN_HPD_DISABLE_REG, &val);
|
regmap_read(pdata->regmap, SN_HPD_DISABLE_REG, &val);
|
||||||
pm_runtime_put_autosuspend(pdata->dev);
|
|
||||||
|
|
||||||
return val & HPD_DEBOUNCED_STATE ? connector_status_connected
|
return val & HPD_DEBOUNCED_STATE ? connector_status_connected
|
||||||
: connector_status_disconnected;
|
: connector_status_disconnected;
|
||||||
@@ -1217,6 +1200,35 @@ static struct edid *ti_sn_bridge_get_edid(struct drm_bridge *bridge,
|
|||||||
return drm_get_edid(connector, &pdata->aux.ddc);
|
return drm_get_edid(connector, &pdata->aux.ddc);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void ti_sn65dsi86_debugfs_init(struct drm_bridge *bridge, struct dentry *root)
|
||||||
|
{
|
||||||
|
struct ti_sn65dsi86 *pdata = bridge_to_ti_sn65dsi86(bridge);
|
||||||
|
struct dentry *debugfs;
|
||||||
|
|
||||||
|
debugfs = debugfs_create_dir(dev_name(pdata->dev), root);
|
||||||
|
debugfs_create_file("status", 0600, debugfs, pdata, &status_fops);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void ti_sn_bridge_hpd_enable(struct drm_bridge *bridge)
|
||||||
|
{
|
||||||
|
struct ti_sn65dsi86 *pdata = bridge_to_ti_sn65dsi86(bridge);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Device needs to be powered on before reading the HPD state
|
||||||
|
* for reliable hpd detection in ti_sn_bridge_detect() due to
|
||||||
|
* the high debounce time.
|
||||||
|
*/
|
||||||
|
|
||||||
|
pm_runtime_get_sync(pdata->dev);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void ti_sn_bridge_hpd_disable(struct drm_bridge *bridge)
|
||||||
|
{
|
||||||
|
struct ti_sn65dsi86 *pdata = bridge_to_ti_sn65dsi86(bridge);
|
||||||
|
|
||||||
|
pm_runtime_put_autosuspend(pdata->dev);
|
||||||
|
}
|
||||||
|
|
||||||
static const struct drm_bridge_funcs ti_sn_bridge_funcs = {
|
static const struct drm_bridge_funcs ti_sn_bridge_funcs = {
|
||||||
.attach = ti_sn_bridge_attach,
|
.attach = ti_sn_bridge_attach,
|
||||||
.detach = ti_sn_bridge_detach,
|
.detach = ti_sn_bridge_detach,
|
||||||
@@ -1230,6 +1242,9 @@ static const struct drm_bridge_funcs ti_sn_bridge_funcs = {
|
|||||||
.atomic_reset = drm_atomic_helper_bridge_reset,
|
.atomic_reset = drm_atomic_helper_bridge_reset,
|
||||||
.atomic_duplicate_state = drm_atomic_helper_bridge_duplicate_state,
|
.atomic_duplicate_state = drm_atomic_helper_bridge_duplicate_state,
|
||||||
.atomic_destroy_state = drm_atomic_helper_bridge_destroy_state,
|
.atomic_destroy_state = drm_atomic_helper_bridge_destroy_state,
|
||||||
|
.debugfs_init = ti_sn65dsi86_debugfs_init,
|
||||||
|
.hpd_enable = ti_sn_bridge_hpd_enable,
|
||||||
|
.hpd_disable = ti_sn_bridge_hpd_disable,
|
||||||
};
|
};
|
||||||
|
|
||||||
static void ti_sn_bridge_parse_lanes(struct ti_sn65dsi86 *pdata,
|
static void ti_sn_bridge_parse_lanes(struct ti_sn65dsi86 *pdata,
|
||||||
@@ -1318,8 +1333,26 @@ static int ti_sn_bridge_probe(struct auxiliary_device *adev,
|
|||||||
pdata->bridge.type = pdata->next_bridge->type == DRM_MODE_CONNECTOR_DisplayPort
|
pdata->bridge.type = pdata->next_bridge->type == DRM_MODE_CONNECTOR_DisplayPort
|
||||||
? DRM_MODE_CONNECTOR_DisplayPort : DRM_MODE_CONNECTOR_eDP;
|
? DRM_MODE_CONNECTOR_DisplayPort : DRM_MODE_CONNECTOR_eDP;
|
||||||
|
|
||||||
if (pdata->bridge.type == DRM_MODE_CONNECTOR_DisplayPort)
|
if (pdata->bridge.type == DRM_MODE_CONNECTOR_DisplayPort) {
|
||||||
pdata->bridge.ops = DRM_BRIDGE_OP_EDID | DRM_BRIDGE_OP_DETECT;
|
pdata->bridge.ops = DRM_BRIDGE_OP_EDID | DRM_BRIDGE_OP_DETECT |
|
||||||
|
DRM_BRIDGE_OP_HPD;
|
||||||
|
/*
|
||||||
|
* If comms were already enabled they would have been enabled
|
||||||
|
* with the wrong value of HPD_DISABLE. Update it now. Comms
|
||||||
|
* could be enabled if anyone is holding a pm_runtime reference
|
||||||
|
* (like if a GPIO is in use). Note that in most cases nobody
|
||||||
|
* is doing AUX channel xfers before the bridge is added so
|
||||||
|
* HPD doesn't _really_ matter then. The only exception is in
|
||||||
|
* the eDP case where the panel wants to read the EDID before
|
||||||
|
* the bridge is added. We always consistently have HPD disabled
|
||||||
|
* for eDP.
|
||||||
|
*/
|
||||||
|
mutex_lock(&pdata->comms_mutex);
|
||||||
|
if (pdata->comms_enabled)
|
||||||
|
regmap_update_bits(pdata->regmap, SN_HPD_DISABLE_REG,
|
||||||
|
HPD_DISABLE, 0);
|
||||||
|
mutex_unlock(&pdata->comms_mutex);
|
||||||
|
};
|
||||||
|
|
||||||
drm_bridge_add(&pdata->bridge);
|
drm_bridge_add(&pdata->bridge);
|
||||||
|
|
||||||
@@ -1935,8 +1968,6 @@ static int ti_sn65dsi86_probe(struct i2c_client *client)
|
|||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
ti_sn65dsi86_debugfs_init(pdata);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Break ourselves up into a collection of aux devices. The only real
|
* Break ourselves up into a collection of aux devices. The only real
|
||||||
* motiviation here is to solve the chicken-and-egg problem of probe
|
* motiviation here is to solve the chicken-and-egg problem of probe
|
||||||
|
@@ -34,6 +34,7 @@ static enum drm_gpu_sched_stat etnaviv_sched_timedout_job(struct drm_sched_job
|
|||||||
*sched_job)
|
*sched_job)
|
||||||
{
|
{
|
||||||
struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
|
struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
|
||||||
|
struct drm_gpu_scheduler *sched = sched_job->sched;
|
||||||
struct etnaviv_gpu *gpu = submit->gpu;
|
struct etnaviv_gpu *gpu = submit->gpu;
|
||||||
u32 dma_addr;
|
u32 dma_addr;
|
||||||
int change;
|
int change;
|
||||||
@@ -76,7 +77,9 @@ static enum drm_gpu_sched_stat etnaviv_sched_timedout_job(struct drm_sched_job
|
|||||||
return DRM_GPU_SCHED_STAT_NOMINAL;
|
return DRM_GPU_SCHED_STAT_NOMINAL;
|
||||||
|
|
||||||
out_no_timeout:
|
out_no_timeout:
|
||||||
list_add(&sched_job->list, &sched_job->sched->pending_list);
|
spin_lock(&sched->job_list_lock);
|
||||||
|
list_add(&sched_job->list, &sched->pending_list);
|
||||||
|
spin_unlock(&sched->job_list_lock);
|
||||||
return DRM_GPU_SCHED_STAT_NOMINAL;
|
return DRM_GPU_SCHED_STAT_NOMINAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -2013,7 +2013,7 @@ static int eb_capture_stage(struct i915_execbuffer *eb)
|
|||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (i915_gem_context_is_recoverable(eb->gem_context) &&
|
if (i915_gem_context_is_recoverable(eb->gem_context) &&
|
||||||
(IS_DGFX(eb->i915) || GRAPHICS_VER_FULL(eb->i915) > IP_VER(12, 0)))
|
GRAPHICS_VER_FULL(eb->i915) > IP_VER(12, 10))
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
for_each_batch_create_order(eb, j) {
|
for_each_batch_create_order(eb, j) {
|
||||||
|
@@ -101,7 +101,7 @@ static unsigned int config_bit(const u64 config)
|
|||||||
return other_bit(config);
|
return other_bit(config);
|
||||||
}
|
}
|
||||||
|
|
||||||
static u32 config_mask(const u64 config)
|
static __always_inline u32 config_mask(const u64 config)
|
||||||
{
|
{
|
||||||
unsigned int bit = config_bit(config);
|
unsigned int bit = config_bit(config);
|
||||||
|
|
||||||
|
@@ -156,6 +156,7 @@ void msm_devfreq_init(struct msm_gpu *gpu)
|
|||||||
priv->gpu_devfreq_config.downdifferential = 10;
|
priv->gpu_devfreq_config.downdifferential = 10;
|
||||||
|
|
||||||
mutex_init(&df->lock);
|
mutex_init(&df->lock);
|
||||||
|
df->suspended = true;
|
||||||
|
|
||||||
ret = dev_pm_qos_add_request(&gpu->pdev->dev, &df->boost_freq,
|
ret = dev_pm_qos_add_request(&gpu->pdev->dev, &df->boost_freq,
|
||||||
DEV_PM_QOS_MIN_FREQUENCY, 0);
|
DEV_PM_QOS_MIN_FREQUENCY, 0);
|
||||||
|
@@ -167,6 +167,7 @@ static void drm_sched_entity_kill_jobs_work(struct work_struct *wrk)
|
|||||||
{
|
{
|
||||||
struct drm_sched_job *job = container_of(wrk, typeof(*job), work);
|
struct drm_sched_job *job = container_of(wrk, typeof(*job), work);
|
||||||
|
|
||||||
|
drm_sched_fence_scheduled(job->s_fence, NULL);
|
||||||
drm_sched_fence_finished(job->s_fence, -ESRCH);
|
drm_sched_fence_finished(job->s_fence, -ESRCH);
|
||||||
WARN_ON(job->s_fence->parent);
|
WARN_ON(job->s_fence->parent);
|
||||||
job->sched->ops->free_job(job);
|
job->sched->ops->free_job(job);
|
||||||
|
@@ -1320,10 +1320,16 @@ static struct drm_plane *tegra_dc_add_shared_planes(struct drm_device *drm,
|
|||||||
if (wgrp->dc == dc->pipe) {
|
if (wgrp->dc == dc->pipe) {
|
||||||
for (j = 0; j < wgrp->num_windows; j++) {
|
for (j = 0; j < wgrp->num_windows; j++) {
|
||||||
unsigned int index = wgrp->windows[j];
|
unsigned int index = wgrp->windows[j];
|
||||||
|
enum drm_plane_type type;
|
||||||
|
|
||||||
|
if (primary)
|
||||||
|
type = DRM_PLANE_TYPE_OVERLAY;
|
||||||
|
else
|
||||||
|
type = DRM_PLANE_TYPE_PRIMARY;
|
||||||
|
|
||||||
plane = tegra_shared_plane_create(drm, dc,
|
plane = tegra_shared_plane_create(drm, dc,
|
||||||
wgrp->index,
|
wgrp->index,
|
||||||
index);
|
index, type);
|
||||||
if (IS_ERR(plane))
|
if (IS_ERR(plane))
|
||||||
return plane;
|
return plane;
|
||||||
|
|
||||||
@@ -1331,13 +1337,11 @@ static struct drm_plane *tegra_dc_add_shared_planes(struct drm_device *drm,
|
|||||||
* Choose the first shared plane owned by this
|
* Choose the first shared plane owned by this
|
||||||
* head as the primary plane.
|
* head as the primary plane.
|
||||||
*/
|
*/
|
||||||
if (!primary) {
|
if (!primary)
|
||||||
plane->type = DRM_PLANE_TYPE_PRIMARY;
|
|
||||||
primary = plane;
|
primary = plane;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
return primary;
|
return primary;
|
||||||
}
|
}
|
||||||
@@ -1388,7 +1392,10 @@ static void tegra_crtc_reset(struct drm_crtc *crtc)
|
|||||||
if (crtc->state)
|
if (crtc->state)
|
||||||
tegra_crtc_atomic_destroy_state(crtc, crtc->state);
|
tegra_crtc_atomic_destroy_state(crtc, crtc->state);
|
||||||
|
|
||||||
|
if (state)
|
||||||
__drm_atomic_helper_crtc_reset(crtc, &state->base);
|
__drm_atomic_helper_crtc_reset(crtc, &state->base);
|
||||||
|
else
|
||||||
|
__drm_atomic_helper_crtc_reset(crtc, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct drm_crtc_state *
|
static struct drm_crtc_state *
|
||||||
|
@@ -756,9 +756,9 @@ static const struct drm_plane_helper_funcs tegra_shared_plane_helper_funcs = {
|
|||||||
struct drm_plane *tegra_shared_plane_create(struct drm_device *drm,
|
struct drm_plane *tegra_shared_plane_create(struct drm_device *drm,
|
||||||
struct tegra_dc *dc,
|
struct tegra_dc *dc,
|
||||||
unsigned int wgrp,
|
unsigned int wgrp,
|
||||||
unsigned int index)
|
unsigned int index,
|
||||||
|
enum drm_plane_type type)
|
||||||
{
|
{
|
||||||
enum drm_plane_type type = DRM_PLANE_TYPE_OVERLAY;
|
|
||||||
struct tegra_drm *tegra = drm->dev_private;
|
struct tegra_drm *tegra = drm->dev_private;
|
||||||
struct tegra_display_hub *hub = tegra->hub;
|
struct tegra_display_hub *hub = tegra->hub;
|
||||||
struct tegra_shared_plane *plane;
|
struct tegra_shared_plane *plane;
|
||||||
|
@@ -80,7 +80,8 @@ void tegra_display_hub_cleanup(struct tegra_display_hub *hub);
|
|||||||
struct drm_plane *tegra_shared_plane_create(struct drm_device *drm,
|
struct drm_plane *tegra_shared_plane_create(struct drm_device *drm,
|
||||||
struct tegra_dc *dc,
|
struct tegra_dc *dc,
|
||||||
unsigned int wgrp,
|
unsigned int wgrp,
|
||||||
unsigned int index);
|
unsigned int index,
|
||||||
|
enum drm_plane_type type);
|
||||||
|
|
||||||
int tegra_display_hub_atomic_check(struct drm_device *drm,
|
int tegra_display_hub_atomic_check(struct drm_device *drm,
|
||||||
struct drm_atomic_state *state);
|
struct drm_atomic_state *state);
|
||||||
|
@@ -318,7 +318,6 @@ static void cirrus_pitch_set(struct cirrus_device *cirrus, unsigned int pitch)
|
|||||||
/* Enable extended blanking and pitch bits, and enable full memory */
|
/* Enable extended blanking and pitch bits, and enable full memory */
|
||||||
cr1b = 0x22;
|
cr1b = 0x22;
|
||||||
cr1b |= (pitch >> 7) & 0x10;
|
cr1b |= (pitch >> 7) & 0x10;
|
||||||
cr1b |= (pitch >> 6) & 0x40;
|
|
||||||
wreg_crt(cirrus, 0x1b, cr1b);
|
wreg_crt(cirrus, 0x1b, cr1b);
|
||||||
|
|
||||||
cirrus_set_start_address(cirrus, 0);
|
cirrus_set_start_address(cirrus, 0);
|
||||||
|
@@ -126,9 +126,9 @@ static void udl_usb_disconnect(struct usb_interface *interface)
|
|||||||
{
|
{
|
||||||
struct drm_device *dev = usb_get_intfdata(interface);
|
struct drm_device *dev = usb_get_intfdata(interface);
|
||||||
|
|
||||||
|
drm_dev_unplug(dev);
|
||||||
drm_kms_helper_poll_fini(dev);
|
drm_kms_helper_poll_fini(dev);
|
||||||
udl_drop_usb(dev);
|
udl_drop_usb(dev);
|
||||||
drm_dev_unplug(dev);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@@ -529,11 +529,14 @@ static void lenovo_features_set_cptkbd(struct hid_device *hdev)
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* Tell the keyboard a driver understands it, and turn F7, F9, F11 into
|
* Tell the keyboard a driver understands it, and turn F7, F9, F11 into
|
||||||
* regular keys
|
* regular keys (Compact only)
|
||||||
*/
|
*/
|
||||||
|
if (hdev->product == USB_DEVICE_ID_LENOVO_CUSBKBD ||
|
||||||
|
hdev->product == USB_DEVICE_ID_LENOVO_CBTKBD) {
|
||||||
ret = lenovo_send_cmd_cptkbd(hdev, 0x01, 0x03);
|
ret = lenovo_send_cmd_cptkbd(hdev, 0x01, 0x03);
|
||||||
if (ret)
|
if (ret)
|
||||||
hid_warn(hdev, "Failed to switch F7/9/11 mode: %d\n", ret);
|
hid_warn(hdev, "Failed to switch F7/9/11 mode: %d\n", ret);
|
||||||
|
}
|
||||||
|
|
||||||
/* Switch middle button to native mode */
|
/* Switch middle button to native mode */
|
||||||
ret = lenovo_send_cmd_cptkbd(hdev, 0x09, 0x01);
|
ret = lenovo_send_cmd_cptkbd(hdev, 0x09, 0x01);
|
||||||
|
@@ -2012,14 +2012,18 @@ static int wacom_initialize_remotes(struct wacom *wacom)
|
|||||||
|
|
||||||
remote->remote_dir = kobject_create_and_add("wacom_remote",
|
remote->remote_dir = kobject_create_and_add("wacom_remote",
|
||||||
&wacom->hdev->dev.kobj);
|
&wacom->hdev->dev.kobj);
|
||||||
if (!remote->remote_dir)
|
if (!remote->remote_dir) {
|
||||||
|
kfifo_free(&remote->remote_fifo);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
error = sysfs_create_files(remote->remote_dir, remote_unpair_attrs);
|
error = sysfs_create_files(remote->remote_dir, remote_unpair_attrs);
|
||||||
|
|
||||||
if (error) {
|
if (error) {
|
||||||
hid_err(wacom->hdev,
|
hid_err(wacom->hdev,
|
||||||
"cannot create sysfs group err: %d\n", error);
|
"cannot create sysfs group err: %d\n", error);
|
||||||
|
kfifo_free(&remote->remote_fifo);
|
||||||
|
kobject_put(remote->remote_dir);
|
||||||
return error;
|
return error;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -120,7 +120,9 @@ const struct vmbus_device vmbus_devs[] = {
|
|||||||
},
|
},
|
||||||
|
|
||||||
/* File copy */
|
/* File copy */
|
||||||
{ .dev_type = HV_FCOPY,
|
/* fcopy always uses 16KB ring buffer size and is working well for last many years */
|
||||||
|
{ .pref_ring_size = 0x4000,
|
||||||
|
.dev_type = HV_FCOPY,
|
||||||
HV_FCOPY_GUID,
|
HV_FCOPY_GUID,
|
||||||
.perf_device = false,
|
.perf_device = false,
|
||||||
.allowed_in_isolated = false,
|
.allowed_in_isolated = false,
|
||||||
@@ -140,12 +142,19 @@ const struct vmbus_device vmbus_devs[] = {
|
|||||||
.allowed_in_isolated = false,
|
.allowed_in_isolated = false,
|
||||||
},
|
},
|
||||||
|
|
||||||
/* Unknown GUID */
|
/*
|
||||||
{ .dev_type = HV_UNKNOWN,
|
* Unknown GUID
|
||||||
|
* 64 KB ring buffer + 4 KB header should be sufficient size for any Hyper-V device apart
|
||||||
|
* from HV_NIC and HV_SCSI. This case avoid the fallback for unknown devices to allocate
|
||||||
|
* much bigger (2 MB) of ring size.
|
||||||
|
*/
|
||||||
|
{ .pref_ring_size = 0x11000,
|
||||||
|
.dev_type = HV_UNKNOWN,
|
||||||
.perf_device = false,
|
.perf_device = false,
|
||||||
.allowed_in_isolated = false,
|
.allowed_in_isolated = false,
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
EXPORT_SYMBOL_GPL(vmbus_devs);
|
||||||
|
|
||||||
static const struct {
|
static const struct {
|
||||||
guid_t guid;
|
guid_t guid;
|
||||||
|
@@ -419,6 +419,11 @@ static inline bool hv_is_perf_channel(struct vmbus_channel *channel)
|
|||||||
return vmbus_devs[channel->device_id].perf_device;
|
return vmbus_devs[channel->device_id].perf_device;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline size_t hv_dev_ring_size(struct vmbus_channel *channel)
|
||||||
|
{
|
||||||
|
return vmbus_devs[channel->device_id].pref_ring_size;
|
||||||
|
}
|
||||||
|
|
||||||
static inline bool hv_is_allocated_cpu(unsigned int cpu)
|
static inline bool hv_is_allocated_cpu(unsigned int cpu)
|
||||||
{
|
{
|
||||||
struct vmbus_channel *channel, *sc;
|
struct vmbus_channel *channel, *sc;
|
||||||
|
@@ -34,16 +34,21 @@ enum chips { max34440, max34441, max34446, max34451, max34460, max34461 };
|
|||||||
/*
|
/*
|
||||||
* The whole max344* family have IOUT_OC_WARN_LIMIT and IOUT_OC_FAULT_LIMIT
|
* The whole max344* family have IOUT_OC_WARN_LIMIT and IOUT_OC_FAULT_LIMIT
|
||||||
* swapped from the standard pmbus spec addresses.
|
* swapped from the standard pmbus spec addresses.
|
||||||
|
* For max34451, version MAX34451ETNA6+ and later has this issue fixed.
|
||||||
*/
|
*/
|
||||||
#define MAX34440_IOUT_OC_WARN_LIMIT 0x46
|
#define MAX34440_IOUT_OC_WARN_LIMIT 0x46
|
||||||
#define MAX34440_IOUT_OC_FAULT_LIMIT 0x4A
|
#define MAX34440_IOUT_OC_FAULT_LIMIT 0x4A
|
||||||
|
|
||||||
|
#define MAX34451ETNA6_MFR_REV 0x0012
|
||||||
|
|
||||||
#define MAX34451_MFR_CHANNEL_CONFIG 0xe4
|
#define MAX34451_MFR_CHANNEL_CONFIG 0xe4
|
||||||
#define MAX34451_MFR_CHANNEL_CONFIG_SEL_MASK 0x3f
|
#define MAX34451_MFR_CHANNEL_CONFIG_SEL_MASK 0x3f
|
||||||
|
|
||||||
struct max34440_data {
|
struct max34440_data {
|
||||||
int id;
|
int id;
|
||||||
struct pmbus_driver_info info;
|
struct pmbus_driver_info info;
|
||||||
|
u8 iout_oc_warn_limit;
|
||||||
|
u8 iout_oc_fault_limit;
|
||||||
};
|
};
|
||||||
|
|
||||||
#define to_max34440_data(x) container_of(x, struct max34440_data, info)
|
#define to_max34440_data(x) container_of(x, struct max34440_data, info)
|
||||||
@@ -60,11 +65,11 @@ static int max34440_read_word_data(struct i2c_client *client, int page,
|
|||||||
switch (reg) {
|
switch (reg) {
|
||||||
case PMBUS_IOUT_OC_FAULT_LIMIT:
|
case PMBUS_IOUT_OC_FAULT_LIMIT:
|
||||||
ret = pmbus_read_word_data(client, page, phase,
|
ret = pmbus_read_word_data(client, page, phase,
|
||||||
MAX34440_IOUT_OC_FAULT_LIMIT);
|
data->iout_oc_fault_limit);
|
||||||
break;
|
break;
|
||||||
case PMBUS_IOUT_OC_WARN_LIMIT:
|
case PMBUS_IOUT_OC_WARN_LIMIT:
|
||||||
ret = pmbus_read_word_data(client, page, phase,
|
ret = pmbus_read_word_data(client, page, phase,
|
||||||
MAX34440_IOUT_OC_WARN_LIMIT);
|
data->iout_oc_warn_limit);
|
||||||
break;
|
break;
|
||||||
case PMBUS_VIRT_READ_VOUT_MIN:
|
case PMBUS_VIRT_READ_VOUT_MIN:
|
||||||
ret = pmbus_read_word_data(client, page, phase,
|
ret = pmbus_read_word_data(client, page, phase,
|
||||||
@@ -133,11 +138,11 @@ static int max34440_write_word_data(struct i2c_client *client, int page,
|
|||||||
|
|
||||||
switch (reg) {
|
switch (reg) {
|
||||||
case PMBUS_IOUT_OC_FAULT_LIMIT:
|
case PMBUS_IOUT_OC_FAULT_LIMIT:
|
||||||
ret = pmbus_write_word_data(client, page, MAX34440_IOUT_OC_FAULT_LIMIT,
|
ret = pmbus_write_word_data(client, page, data->iout_oc_fault_limit,
|
||||||
word);
|
word);
|
||||||
break;
|
break;
|
||||||
case PMBUS_IOUT_OC_WARN_LIMIT:
|
case PMBUS_IOUT_OC_WARN_LIMIT:
|
||||||
ret = pmbus_write_word_data(client, page, MAX34440_IOUT_OC_WARN_LIMIT,
|
ret = pmbus_write_word_data(client, page, data->iout_oc_warn_limit,
|
||||||
word);
|
word);
|
||||||
break;
|
break;
|
||||||
case PMBUS_VIRT_RESET_POUT_HISTORY:
|
case PMBUS_VIRT_RESET_POUT_HISTORY:
|
||||||
@@ -235,6 +240,25 @@ static int max34451_set_supported_funcs(struct i2c_client *client,
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
int page, rv;
|
int page, rv;
|
||||||
|
bool max34451_na6 = false;
|
||||||
|
|
||||||
|
rv = i2c_smbus_read_word_data(client, PMBUS_MFR_REVISION);
|
||||||
|
if (rv < 0)
|
||||||
|
return rv;
|
||||||
|
|
||||||
|
if (rv >= MAX34451ETNA6_MFR_REV) {
|
||||||
|
max34451_na6 = true;
|
||||||
|
data->info.format[PSC_VOLTAGE_IN] = direct;
|
||||||
|
data->info.format[PSC_CURRENT_IN] = direct;
|
||||||
|
data->info.m[PSC_VOLTAGE_IN] = 1;
|
||||||
|
data->info.b[PSC_VOLTAGE_IN] = 0;
|
||||||
|
data->info.R[PSC_VOLTAGE_IN] = 3;
|
||||||
|
data->info.m[PSC_CURRENT_IN] = 1;
|
||||||
|
data->info.b[PSC_CURRENT_IN] = 0;
|
||||||
|
data->info.R[PSC_CURRENT_IN] = 2;
|
||||||
|
data->iout_oc_fault_limit = PMBUS_IOUT_OC_FAULT_LIMIT;
|
||||||
|
data->iout_oc_warn_limit = PMBUS_IOUT_OC_WARN_LIMIT;
|
||||||
|
}
|
||||||
|
|
||||||
for (page = 0; page < 16; page++) {
|
for (page = 0; page < 16; page++) {
|
||||||
rv = i2c_smbus_write_byte_data(client, PMBUS_PAGE, page);
|
rv = i2c_smbus_write_byte_data(client, PMBUS_PAGE, page);
|
||||||
@@ -251,16 +275,30 @@ static int max34451_set_supported_funcs(struct i2c_client *client,
|
|||||||
case 0x20:
|
case 0x20:
|
||||||
data->info.func[page] = PMBUS_HAVE_VOUT |
|
data->info.func[page] = PMBUS_HAVE_VOUT |
|
||||||
PMBUS_HAVE_STATUS_VOUT;
|
PMBUS_HAVE_STATUS_VOUT;
|
||||||
|
|
||||||
|
if (max34451_na6)
|
||||||
|
data->info.func[page] |= PMBUS_HAVE_VIN |
|
||||||
|
PMBUS_HAVE_STATUS_INPUT;
|
||||||
break;
|
break;
|
||||||
case 0x21:
|
case 0x21:
|
||||||
data->info.func[page] = PMBUS_HAVE_VOUT;
|
data->info.func[page] = PMBUS_HAVE_VOUT;
|
||||||
|
|
||||||
|
if (max34451_na6)
|
||||||
|
data->info.func[page] |= PMBUS_HAVE_VIN;
|
||||||
break;
|
break;
|
||||||
case 0x22:
|
case 0x22:
|
||||||
data->info.func[page] = PMBUS_HAVE_IOUT |
|
data->info.func[page] = PMBUS_HAVE_IOUT |
|
||||||
PMBUS_HAVE_STATUS_IOUT;
|
PMBUS_HAVE_STATUS_IOUT;
|
||||||
|
|
||||||
|
if (max34451_na6)
|
||||||
|
data->info.func[page] |= PMBUS_HAVE_IIN |
|
||||||
|
PMBUS_HAVE_STATUS_INPUT;
|
||||||
break;
|
break;
|
||||||
case 0x23:
|
case 0x23:
|
||||||
data->info.func[page] = PMBUS_HAVE_IOUT;
|
data->info.func[page] = PMBUS_HAVE_IOUT;
|
||||||
|
|
||||||
|
if (max34451_na6)
|
||||||
|
data->info.func[page] |= PMBUS_HAVE_IIN;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
@@ -494,6 +532,8 @@ static int max34440_probe(struct i2c_client *client)
|
|||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
data->id = i2c_match_id(max34440_id, client)->driver_data;
|
data->id = i2c_match_id(max34440_id, client)->driver_data;
|
||||||
data->info = max34440_info[data->id];
|
data->info = max34440_info[data->id];
|
||||||
|
data->iout_oc_fault_limit = MAX34440_IOUT_OC_FAULT_LIMIT;
|
||||||
|
data->iout_oc_warn_limit = MAX34440_IOUT_OC_WARN_LIMIT;
|
||||||
|
|
||||||
if (data->id == max34451) {
|
if (data->id == max34451) {
|
||||||
rv = max34451_set_supported_funcs(client, data);
|
rv = max34451_set_supported_funcs(client, data);
|
||||||
|
@@ -135,7 +135,8 @@ coresight_find_out_connection(struct coresight_device *src_dev,
|
|||||||
|
|
||||||
static inline u32 coresight_read_claim_tags(struct coresight_device *csdev)
|
static inline u32 coresight_read_claim_tags(struct coresight_device *csdev)
|
||||||
{
|
{
|
||||||
return csdev_access_relaxed_read32(&csdev->access, CORESIGHT_CLAIMCLR);
|
return FIELD_GET(CORESIGHT_CLAIM_MASK,
|
||||||
|
csdev_access_relaxed_read32(&csdev->access, CORESIGHT_CLAIMCLR));
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline bool coresight_is_claimed_self_hosted(struct coresight_device *csdev)
|
static inline bool coresight_is_claimed_self_hosted(struct coresight_device *csdev)
|
||||||
|
@@ -32,6 +32,7 @@
|
|||||||
* Coresight device CLAIM protocol.
|
* Coresight device CLAIM protocol.
|
||||||
* See PSCI - ARM DEN 0022D, Section: 6.8.1 Debug and Trace save and restore.
|
* See PSCI - ARM DEN 0022D, Section: 6.8.1 Debug and Trace save and restore.
|
||||||
*/
|
*/
|
||||||
|
#define CORESIGHT_CLAIM_MASK GENMASK(1, 0)
|
||||||
#define CORESIGHT_CLAIM_SELF_HOSTED BIT(1)
|
#define CORESIGHT_CLAIM_SELF_HOSTED BIT(1)
|
||||||
|
|
||||||
#define TIMEOUT_US 100
|
#define TIMEOUT_US 100
|
||||||
|
@@ -111,6 +111,11 @@ static u32 osif_func(struct i2c_adapter *adapter)
|
|||||||
return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL;
|
return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* prevent invalid 0-length usb_control_msg */
|
||||||
|
static const struct i2c_adapter_quirks osif_quirks = {
|
||||||
|
.flags = I2C_AQ_NO_ZERO_LEN_READ,
|
||||||
|
};
|
||||||
|
|
||||||
static const struct i2c_algorithm osif_algorithm = {
|
static const struct i2c_algorithm osif_algorithm = {
|
||||||
.master_xfer = osif_xfer,
|
.master_xfer = osif_xfer,
|
||||||
.functionality = osif_func,
|
.functionality = osif_func,
|
||||||
@@ -143,6 +148,7 @@ static int osif_probe(struct usb_interface *interface,
|
|||||||
|
|
||||||
priv->adapter.owner = THIS_MODULE;
|
priv->adapter.owner = THIS_MODULE;
|
||||||
priv->adapter.class = I2C_CLASS_HWMON;
|
priv->adapter.class = I2C_CLASS_HWMON;
|
||||||
|
priv->adapter.quirks = &osif_quirks;
|
||||||
priv->adapter.algo = &osif_algorithm;
|
priv->adapter.algo = &osif_algorithm;
|
||||||
priv->adapter.algo_data = priv;
|
priv->adapter.algo_data = priv;
|
||||||
snprintf(priv->adapter.name, sizeof(priv->adapter.name),
|
snprintf(priv->adapter.name, sizeof(priv->adapter.name),
|
||||||
|
@@ -140,6 +140,11 @@ out:
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* prevent invalid 0-length usb_control_msg */
|
||||||
|
static const struct i2c_adapter_quirks usb_quirks = {
|
||||||
|
.flags = I2C_AQ_NO_ZERO_LEN_READ,
|
||||||
|
};
|
||||||
|
|
||||||
/* This is the actual algorithm we define */
|
/* This is the actual algorithm we define */
|
||||||
static const struct i2c_algorithm usb_algorithm = {
|
static const struct i2c_algorithm usb_algorithm = {
|
||||||
.master_xfer = usb_xfer,
|
.master_xfer = usb_xfer,
|
||||||
@@ -248,6 +253,7 @@ static int i2c_tiny_usb_probe(struct usb_interface *interface,
|
|||||||
/* setup i2c adapter description */
|
/* setup i2c adapter description */
|
||||||
dev->adapter.owner = THIS_MODULE;
|
dev->adapter.owner = THIS_MODULE;
|
||||||
dev->adapter.class = I2C_CLASS_HWMON;
|
dev->adapter.class = I2C_CLASS_HWMON;
|
||||||
|
dev->adapter.quirks = &usb_quirks;
|
||||||
dev->adapter.algo = &usb_algorithm;
|
dev->adapter.algo = &usb_algorithm;
|
||||||
dev->adapter.algo_data = dev;
|
dev->adapter.algo_data = dev;
|
||||||
snprintf(dev->adapter.name, sizeof(dev->adapter.name),
|
snprintf(dev->adapter.name, sizeof(dev->adapter.name),
|
||||||
|
@@ -476,6 +476,10 @@ static irqreturn_t ad_sd_trigger_handler(int irq, void *p)
|
|||||||
* byte set to zero. */
|
* byte set to zero. */
|
||||||
ad_sd_read_reg_raw(sigma_delta, data_reg, transfer_size, &data[1]);
|
ad_sd_read_reg_raw(sigma_delta, data_reg, transfer_size, &data[1]);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
default:
|
||||||
|
dev_err_ratelimited(&indio_dev->dev, "Unsupported reg_size: %u\n", reg_size);
|
||||||
|
goto irq_handled;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@@ -582,7 +582,7 @@ static int zpa2326_fill_sample_buffer(struct iio_dev *indio_dev,
|
|||||||
struct {
|
struct {
|
||||||
u32 pressure;
|
u32 pressure;
|
||||||
u16 temperature;
|
u16 temperature;
|
||||||
u64 timestamp;
|
aligned_s64 timestamp;
|
||||||
} sample;
|
} sample;
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
|
@@ -61,6 +61,7 @@ static ssize_t multi_intensity_store(struct device *dev,
|
|||||||
for (i = 0; i < mcled_cdev->num_colors; i++)
|
for (i = 0; i < mcled_cdev->num_colors; i++)
|
||||||
mcled_cdev->subled_info[i].intensity = intensity_value[i];
|
mcled_cdev->subled_info[i].intensity = intensity_value[i];
|
||||||
|
|
||||||
|
if (!test_bit(LED_BLINK_SW, &led_cdev->work_flags))
|
||||||
led_set_brightness(led_cdev, led_cdev->brightness);
|
led_set_brightness(led_cdev, led_cdev->brightness);
|
||||||
ret = size;
|
ret = size;
|
||||||
err_out:
|
err_out:
|
||||||
|
@@ -500,8 +500,8 @@ void mbox_free_channel(struct mbox_chan *chan)
|
|||||||
if (chan->txdone_method == TXDONE_BY_ACK)
|
if (chan->txdone_method == TXDONE_BY_ACK)
|
||||||
chan->txdone_method = TXDONE_BY_POLL;
|
chan->txdone_method = TXDONE_BY_POLL;
|
||||||
|
|
||||||
module_put(chan->mbox->dev->driver->owner);
|
|
||||||
spin_unlock_irqrestore(&chan->lock, flags);
|
spin_unlock_irqrestore(&chan->lock, flags);
|
||||||
|
module_put(chan->mbox->dev->driver->owner);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(mbox_free_channel);
|
EXPORT_SYMBOL_GPL(mbox_free_channel);
|
||||||
|
|
||||||
|
@@ -1739,7 +1739,12 @@ static void cache_set_flush(struct closure *cl)
|
|||||||
mutex_unlock(&b->write_lock);
|
mutex_unlock(&b->write_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ca->alloc_thread)
|
/*
|
||||||
|
* If the register_cache_set() call to bch_cache_set_alloc() failed,
|
||||||
|
* ca has not been assigned a value and return error.
|
||||||
|
* So we need check ca is not NULL during bch_cache_set_unregister().
|
||||||
|
*/
|
||||||
|
if (ca && ca->alloc_thread)
|
||||||
kthread_stop(ca->alloc_thread);
|
kthread_stop(ca->alloc_thread);
|
||||||
|
|
||||||
if (c->journal.cur) {
|
if (c->journal.cur) {
|
||||||
|
@@ -2378,7 +2378,7 @@ static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev)
|
|||||||
*/
|
*/
|
||||||
sb_retrieve_failed_devices(sb, failed_devices);
|
sb_retrieve_failed_devices(sb, failed_devices);
|
||||||
rdev_for_each(r, mddev) {
|
rdev_for_each(r, mddev) {
|
||||||
if (test_bit(Journal, &rdev->flags) ||
|
if (test_bit(Journal, &r->flags) ||
|
||||||
!r->sb_page)
|
!r->sb_page)
|
||||||
continue;
|
continue;
|
||||||
sb2 = page_address(r->sb_page);
|
sb2 = page_address(r->sb_page);
|
||||||
|
@@ -589,7 +589,7 @@ static int md_bitmap_new_disk_sb(struct bitmap *bitmap)
|
|||||||
* is a good choice? We choose COUNTER_MAX / 2 arbitrarily.
|
* is a good choice? We choose COUNTER_MAX / 2 arbitrarily.
|
||||||
*/
|
*/
|
||||||
write_behind = bitmap->mddev->bitmap_info.max_write_behind;
|
write_behind = bitmap->mddev->bitmap_info.max_write_behind;
|
||||||
if (write_behind > COUNTER_MAX)
|
if (write_behind > COUNTER_MAX / 2)
|
||||||
write_behind = COUNTER_MAX / 2;
|
write_behind = COUNTER_MAX / 2;
|
||||||
sb->write_behind = cpu_to_le32(write_behind);
|
sb->write_behind = cpu_to_le32(write_behind);
|
||||||
bitmap->mddev->bitmap_info.max_write_behind = write_behind;
|
bitmap->mddev->bitmap_info.max_write_behind = write_behind;
|
||||||
|
@@ -1801,7 +1801,7 @@ static int uvc_ctrl_commit_entity(struct uvc_device *dev,
|
|||||||
unsigned int processed_ctrls = 0;
|
unsigned int processed_ctrls = 0;
|
||||||
struct uvc_control *ctrl;
|
struct uvc_control *ctrl;
|
||||||
unsigned int i;
|
unsigned int i;
|
||||||
int ret;
|
int ret = 0;
|
||||||
|
|
||||||
if (entity == NULL)
|
if (entity == NULL)
|
||||||
return 0;
|
return 0;
|
||||||
@@ -1830,8 +1830,6 @@ static int uvc_ctrl_commit_entity(struct uvc_device *dev,
|
|||||||
dev->intfnum, ctrl->info.selector,
|
dev->intfnum, ctrl->info.selector,
|
||||||
uvc_ctrl_data(ctrl, UVC_CTRL_DATA_CURRENT),
|
uvc_ctrl_data(ctrl, UVC_CTRL_DATA_CURRENT),
|
||||||
ctrl->info.size);
|
ctrl->info.size);
|
||||||
else
|
|
||||||
ret = 0;
|
|
||||||
|
|
||||||
if (!ret)
|
if (!ret)
|
||||||
processed_ctrls++;
|
processed_ctrls++;
|
||||||
@@ -1843,17 +1841,25 @@ static int uvc_ctrl_commit_entity(struct uvc_device *dev,
|
|||||||
|
|
||||||
ctrl->dirty = 0;
|
ctrl->dirty = 0;
|
||||||
|
|
||||||
if (ret < 0) {
|
|
||||||
if (err_ctrl)
|
|
||||||
*err_ctrl = ctrl;
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!rollback && handle &&
|
if (!rollback && handle &&
|
||||||
ctrl->info.flags & UVC_CTRL_FLAG_ASYNCHRONOUS)
|
ctrl->info.flags & UVC_CTRL_FLAG_ASYNCHRONOUS)
|
||||||
uvc_ctrl_set_handle(handle, ctrl, handle);
|
uvc_ctrl_set_handle(handle, ctrl, handle);
|
||||||
|
|
||||||
|
if (ret < 0 && !rollback) {
|
||||||
|
if (err_ctrl)
|
||||||
|
*err_ctrl = ctrl;
|
||||||
|
/*
|
||||||
|
* If we fail to set a control, we need to rollback
|
||||||
|
* the next ones.
|
||||||
|
*/
|
||||||
|
rollback = 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
return processed_ctrls;
|
return processed_ctrls;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1884,7 +1890,8 @@ int __uvc_ctrl_commit(struct uvc_fh *handle, int rollback,
|
|||||||
struct uvc_video_chain *chain = handle->chain;
|
struct uvc_video_chain *chain = handle->chain;
|
||||||
struct uvc_control *err_ctrl;
|
struct uvc_control *err_ctrl;
|
||||||
struct uvc_entity *entity;
|
struct uvc_entity *entity;
|
||||||
int ret = 0;
|
int ret_out = 0;
|
||||||
|
int ret;
|
||||||
|
|
||||||
/* Find the control. */
|
/* Find the control. */
|
||||||
list_for_each_entry(entity, &chain->entities, chain) {
|
list_for_each_entry(entity, &chain->entities, chain) {
|
||||||
@@ -1895,17 +1902,23 @@ int __uvc_ctrl_commit(struct uvc_fh *handle, int rollback,
|
|||||||
ctrls->error_idx =
|
ctrls->error_idx =
|
||||||
uvc_ctrl_find_ctrl_idx(entity, ctrls,
|
uvc_ctrl_find_ctrl_idx(entity, ctrls,
|
||||||
err_ctrl);
|
err_ctrl);
|
||||||
goto done;
|
/*
|
||||||
|
* When we fail to commit an entity, we need to
|
||||||
|
* restore the UVC_CTRL_DATA_BACKUP for all the
|
||||||
|
* controls in the other entities, otherwise our cache
|
||||||
|
* and the hardware will be out of sync.
|
||||||
|
*/
|
||||||
|
rollback = 1;
|
||||||
|
|
||||||
|
ret_out = ret;
|
||||||
} else if (ret > 0 && !rollback) {
|
} else if (ret > 0 && !rollback) {
|
||||||
uvc_ctrl_send_events(handle, entity,
|
uvc_ctrl_send_events(handle, entity,
|
||||||
ctrls->controls, ctrls->count);
|
ctrls->controls, ctrls->count);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = 0;
|
|
||||||
done:
|
|
||||||
mutex_unlock(&chain->ctrl_mutex);
|
mutex_unlock(&chain->ctrl_mutex);
|
||||||
return ret;
|
return ret_out;
|
||||||
}
|
}
|
||||||
|
|
||||||
int uvc_ctrl_get(struct uvc_video_chain *chain,
|
int uvc_ctrl_get(struct uvc_video_chain *chain,
|
||||||
|
@@ -463,6 +463,7 @@ static void max14577_i2c_remove(struct i2c_client *i2c)
|
|||||||
{
|
{
|
||||||
struct max14577 *max14577 = i2c_get_clientdata(i2c);
|
struct max14577 *max14577 = i2c_get_clientdata(i2c);
|
||||||
|
|
||||||
|
device_init_wakeup(max14577->dev, false);
|
||||||
mfd_remove_devices(max14577->dev);
|
mfd_remove_devices(max14577->dev);
|
||||||
regmap_del_irq_chip(max14577->irq, max14577->irq_data);
|
regmap_del_irq_chip(max14577->irq, max14577->irq_data);
|
||||||
if (max14577->dev_type == MAXIM_DEVICE_TYPE_MAX77836)
|
if (max14577->dev_type == MAXIM_DEVICE_TYPE_MAX77836)
|
||||||
|
@@ -260,6 +260,9 @@ static int tps6594_pfsm_probe(struct platform_device *pdev)
|
|||||||
pfsm->miscdev.minor = MISC_DYNAMIC_MINOR;
|
pfsm->miscdev.minor = MISC_DYNAMIC_MINOR;
|
||||||
pfsm->miscdev.name = devm_kasprintf(dev, GFP_KERNEL, "pfsm-%ld-0x%02x",
|
pfsm->miscdev.name = devm_kasprintf(dev, GFP_KERNEL, "pfsm-%ld-0x%02x",
|
||||||
tps->chip_id, tps->reg);
|
tps->chip_id, tps->reg);
|
||||||
|
if (!pfsm->miscdev.name)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
pfsm->miscdev.fops = &tps6594_pfsm_fops;
|
pfsm->miscdev.fops = &tps6594_pfsm_fops;
|
||||||
pfsm->miscdev.parent = dev->parent;
|
pfsm->miscdev.parent = dev->parent;
|
||||||
|
|
||||||
|
@@ -485,7 +485,7 @@ static inline u64 _enetc_rd_reg64(void __iomem *reg)
|
|||||||
tmp = ioread32(reg + 4);
|
tmp = ioread32(reg + 4);
|
||||||
} while (high != tmp);
|
} while (high != tmp);
|
||||||
|
|
||||||
return le64_to_cpu((__le64)high << 32 | low);
|
return (u64)high << 32 | low;
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@@ -2368,7 +2368,7 @@ static int wx_alloc_page_pool(struct wx_ring *rx_ring)
|
|||||||
struct page_pool_params pp_params = {
|
struct page_pool_params pp_params = {
|
||||||
.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV,
|
.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV,
|
||||||
.order = 0,
|
.order = 0,
|
||||||
.pool_size = rx_ring->size,
|
.pool_size = rx_ring->count,
|
||||||
.nid = dev_to_node(rx_ring->dev),
|
.nid = dev_to_node(rx_ring->dev),
|
||||||
.dev = rx_ring->dev,
|
.dev = rx_ring->dev,
|
||||||
.dma_dir = DMA_FROM_DEVICE,
|
.dma_dir = DMA_FROM_DEVICE,
|
||||||
|
@@ -526,16 +526,14 @@ static enum rq_end_io_ret nvme_uring_cmd_end_io(struct request *req,
|
|||||||
pdu->u.result = le64_to_cpu(nvme_req(req)->result.u64);
|
pdu->u.result = le64_to_cpu(nvme_req(req)->result.u64);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* For iopoll, complete it directly.
|
* IOPOLL could potentially complete this request directly, but
|
||||||
* Otherwise, move the completion to task work.
|
* if multiple rings are polling on the same queue, then it's possible
|
||||||
|
* for one ring to find completions for another ring. Punting the
|
||||||
|
* completion via task_work will always direct it to the right
|
||||||
|
* location, rather than potentially complete requests for ringA
|
||||||
|
* under iopoll invocations from ringB.
|
||||||
*/
|
*/
|
||||||
if (blk_rq_is_poll(req)) {
|
|
||||||
WRITE_ONCE(ioucmd->cookie, NULL);
|
|
||||||
nvme_uring_task_cb(ioucmd, IO_URING_F_UNLOCKED);
|
|
||||||
} else {
|
|
||||||
io_uring_cmd_do_in_task_lazy(ioucmd, nvme_uring_task_cb);
|
io_uring_cmd_do_in_task_lazy(ioucmd, nvme_uring_task_cb);
|
||||||
}
|
|
||||||
|
|
||||||
return RQ_END_IO_FREE;
|
return RQ_END_IO_FREE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -748,22 +748,19 @@ static void dw_pcie_link_set_max_link_width(struct dw_pcie *pci, u32 num_lanes)
|
|||||||
/* Set link width speed control register */
|
/* Set link width speed control register */
|
||||||
lwsc = dw_pcie_readl_dbi(pci, PCIE_LINK_WIDTH_SPEED_CONTROL);
|
lwsc = dw_pcie_readl_dbi(pci, PCIE_LINK_WIDTH_SPEED_CONTROL);
|
||||||
lwsc &= ~PORT_LOGIC_LINK_WIDTH_MASK;
|
lwsc &= ~PORT_LOGIC_LINK_WIDTH_MASK;
|
||||||
|
lwsc |= PORT_LOGIC_LINK_WIDTH_1_LANES;
|
||||||
switch (num_lanes) {
|
switch (num_lanes) {
|
||||||
case 1:
|
case 1:
|
||||||
plc |= PORT_LINK_MODE_1_LANES;
|
plc |= PORT_LINK_MODE_1_LANES;
|
||||||
lwsc |= PORT_LOGIC_LINK_WIDTH_1_LANES;
|
|
||||||
break;
|
break;
|
||||||
case 2:
|
case 2:
|
||||||
plc |= PORT_LINK_MODE_2_LANES;
|
plc |= PORT_LINK_MODE_2_LANES;
|
||||||
lwsc |= PORT_LOGIC_LINK_WIDTH_2_LANES;
|
|
||||||
break;
|
break;
|
||||||
case 4:
|
case 4:
|
||||||
plc |= PORT_LINK_MODE_4_LANES;
|
plc |= PORT_LINK_MODE_4_LANES;
|
||||||
lwsc |= PORT_LOGIC_LINK_WIDTH_4_LANES;
|
|
||||||
break;
|
break;
|
||||||
case 8:
|
case 8:
|
||||||
plc |= PORT_LINK_MODE_8_LANES;
|
plc |= PORT_LINK_MODE_8_LANES;
|
||||||
lwsc |= PORT_LOGIC_LINK_WIDTH_8_LANES;
|
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
dev_err(pci->dev, "num-lanes %u: invalid value\n", num_lanes);
|
dev_err(pci->dev, "num-lanes %u: invalid value\n", num_lanes);
|
||||||
|
@@ -585,6 +585,9 @@ static int apple_pcie_setup_port(struct apple_pcie *pcie,
|
|||||||
list_add_tail(&port->entry, &pcie->ports);
|
list_add_tail(&port->entry, &pcie->ports);
|
||||||
init_completion(&pcie->event);
|
init_completion(&pcie->event);
|
||||||
|
|
||||||
|
/* In the success path, we keep a reference to np around */
|
||||||
|
of_node_get(np);
|
||||||
|
|
||||||
ret = apple_pcie_port_register_irqs(port);
|
ret = apple_pcie_port_register_irqs(port);
|
||||||
WARN_ON(ret);
|
WARN_ON(ret);
|
||||||
|
|
||||||
@@ -764,7 +767,6 @@ static int apple_pcie_init(struct pci_config_window *cfg)
|
|||||||
{
|
{
|
||||||
struct device *dev = cfg->parent;
|
struct device *dev = cfg->parent;
|
||||||
struct platform_device *platform = to_platform_device(dev);
|
struct platform_device *platform = to_platform_device(dev);
|
||||||
struct device_node *of_port;
|
|
||||||
struct apple_pcie *pcie;
|
struct apple_pcie *pcie;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
@@ -787,11 +789,10 @@ static int apple_pcie_init(struct pci_config_window *cfg)
|
|||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
for_each_child_of_node(dev->of_node, of_port) {
|
for_each_available_child_of_node_scoped(dev->of_node, of_port) {
|
||||||
ret = apple_pcie_setup_port(pcie, of_port);
|
ret = apple_pcie_setup_port(pcie, of_port);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
dev_err(pcie->dev, "Port %pOF setup fail: %d\n", of_port, ret);
|
dev_err(pcie->dev, "Port %pOF setup fail: %d\n", of_port, ret);
|
||||||
of_node_put(of_port);
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -466,6 +466,7 @@ config LENOVO_YMC
|
|||||||
tristate "Lenovo Yoga Tablet Mode Control"
|
tristate "Lenovo Yoga Tablet Mode Control"
|
||||||
depends on ACPI_WMI
|
depends on ACPI_WMI
|
||||||
depends on INPUT
|
depends on INPUT
|
||||||
|
depends on IDEAPAD_LAPTOP
|
||||||
select INPUT_SPARSEKMAP
|
select INPUT_SPARSEKMAP
|
||||||
help
|
help
|
||||||
This driver maps the Tablet Mode Control switch to SW_TABLET_MODE input
|
This driver maps the Tablet Mode Control switch to SW_TABLET_MODE input
|
||||||
|
@@ -14,6 +14,7 @@
|
|||||||
#include <linux/bitops.h>
|
#include <linux/bitops.h>
|
||||||
#include <linux/bug.h>
|
#include <linux/bug.h>
|
||||||
#include <linux/debugfs.h>
|
#include <linux/debugfs.h>
|
||||||
|
#include <linux/delay.h>
|
||||||
#include <linux/device.h>
|
#include <linux/device.h>
|
||||||
#include <linux/dmi.h>
|
#include <linux/dmi.h>
|
||||||
#include <linux/fb.h>
|
#include <linux/fb.h>
|
||||||
@@ -21,6 +22,7 @@
|
|||||||
#include <linux/init.h>
|
#include <linux/init.h>
|
||||||
#include <linux/input.h>
|
#include <linux/input.h>
|
||||||
#include <linux/input/sparse-keymap.h>
|
#include <linux/input/sparse-keymap.h>
|
||||||
|
#include <linux/jiffies.h>
|
||||||
#include <linux/kernel.h>
|
#include <linux/kernel.h>
|
||||||
#include <linux/leds.h>
|
#include <linux/leds.h>
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
@@ -86,6 +88,34 @@ enum {
|
|||||||
SALS_FNLOCK_OFF = 0xf,
|
SALS_FNLOCK_OFF = 0xf,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum {
|
||||||
|
VPCCMD_R_VPC1 = 0x10,
|
||||||
|
VPCCMD_R_BL_MAX,
|
||||||
|
VPCCMD_R_BL,
|
||||||
|
VPCCMD_W_BL,
|
||||||
|
VPCCMD_R_WIFI,
|
||||||
|
VPCCMD_W_WIFI,
|
||||||
|
VPCCMD_R_BT,
|
||||||
|
VPCCMD_W_BT,
|
||||||
|
VPCCMD_R_BL_POWER,
|
||||||
|
VPCCMD_R_NOVO,
|
||||||
|
VPCCMD_R_VPC2,
|
||||||
|
VPCCMD_R_TOUCHPAD,
|
||||||
|
VPCCMD_W_TOUCHPAD,
|
||||||
|
VPCCMD_R_CAMERA,
|
||||||
|
VPCCMD_W_CAMERA,
|
||||||
|
VPCCMD_R_3G,
|
||||||
|
VPCCMD_W_3G,
|
||||||
|
VPCCMD_R_ODD, /* 0x21 */
|
||||||
|
VPCCMD_W_FAN,
|
||||||
|
VPCCMD_R_RF,
|
||||||
|
VPCCMD_W_RF,
|
||||||
|
VPCCMD_W_YMC = 0x2A,
|
||||||
|
VPCCMD_R_FAN = 0x2B,
|
||||||
|
VPCCMD_R_SPECIAL_BUTTONS = 0x31,
|
||||||
|
VPCCMD_W_BL_POWER = 0x33,
|
||||||
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* These correspond to the number of supported states - 1
|
* These correspond to the number of supported states - 1
|
||||||
* Future keyboard types may need a new system, if there's a collision
|
* Future keyboard types may need a new system, if there's a collision
|
||||||
@@ -145,6 +175,7 @@ struct ideapad_private {
|
|||||||
bool touchpad_ctrl_via_ec : 1;
|
bool touchpad_ctrl_via_ec : 1;
|
||||||
bool ctrl_ps2_aux_port : 1;
|
bool ctrl_ps2_aux_port : 1;
|
||||||
bool usb_charging : 1;
|
bool usb_charging : 1;
|
||||||
|
bool ymc_ec_trigger : 1;
|
||||||
} features;
|
} features;
|
||||||
struct {
|
struct {
|
||||||
bool initialized;
|
bool initialized;
|
||||||
@@ -188,6 +219,12 @@ MODULE_PARM_DESC(touchpad_ctrl_via_ec,
|
|||||||
"Enable registering a 'touchpad' sysfs-attribute which can be used to manually "
|
"Enable registering a 'touchpad' sysfs-attribute which can be used to manually "
|
||||||
"tell the EC to enable/disable the touchpad. This may not work on all models.");
|
"tell the EC to enable/disable the touchpad. This may not work on all models.");
|
||||||
|
|
||||||
|
static bool ymc_ec_trigger __read_mostly;
|
||||||
|
module_param(ymc_ec_trigger, bool, 0444);
|
||||||
|
MODULE_PARM_DESC(ymc_ec_trigger,
|
||||||
|
"Enable EC triggering work-around to force emitting tablet mode events. "
|
||||||
|
"If you need this please report this to: platform-driver-x86@vger.kernel.org");
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* shared data
|
* shared data
|
||||||
*/
|
*/
|
||||||
@@ -227,6 +264,21 @@ static void ideapad_shared_exit(struct ideapad_private *priv)
|
|||||||
/*
|
/*
|
||||||
* ACPI Helpers
|
* ACPI Helpers
|
||||||
*/
|
*/
|
||||||
|
#define IDEAPAD_EC_TIMEOUT 200 /* in ms */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Some models (e.g., ThinkBook since 2024) have a low tolerance for being
|
||||||
|
* polled too frequently. Doing so may break the state machine in the EC,
|
||||||
|
* resulting in a hard shutdown.
|
||||||
|
*
|
||||||
|
* It is also observed that frequent polls may disturb the ongoing operation
|
||||||
|
* and notably delay the availability of EC response.
|
||||||
|
*
|
||||||
|
* These values are used as the delay before the first poll and the interval
|
||||||
|
* between subsequent polls to solve the above issues.
|
||||||
|
*/
|
||||||
|
#define IDEAPAD_EC_POLL_MIN_US 150
|
||||||
|
#define IDEAPAD_EC_POLL_MAX_US 300
|
||||||
|
|
||||||
static int eval_int(acpi_handle handle, const char *name, unsigned long *res)
|
static int eval_int(acpi_handle handle, const char *name, unsigned long *res)
|
||||||
{
|
{
|
||||||
@@ -242,6 +294,29 @@ static int eval_int(acpi_handle handle, const char *name, unsigned long *res)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int eval_int_with_arg(acpi_handle handle, const char *name, unsigned long arg,
|
||||||
|
unsigned long *res)
|
||||||
|
{
|
||||||
|
struct acpi_object_list params;
|
||||||
|
unsigned long long result;
|
||||||
|
union acpi_object in_obj;
|
||||||
|
acpi_status status;
|
||||||
|
|
||||||
|
params.count = 1;
|
||||||
|
params.pointer = &in_obj;
|
||||||
|
in_obj.type = ACPI_TYPE_INTEGER;
|
||||||
|
in_obj.integer.value = arg;
|
||||||
|
|
||||||
|
status = acpi_evaluate_integer(handle, (char *)name, ¶ms, &result);
|
||||||
|
if (ACPI_FAILURE(status))
|
||||||
|
return -EIO;
|
||||||
|
|
||||||
|
if (res)
|
||||||
|
*res = result;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
static int exec_simple_method(acpi_handle handle, const char *name, unsigned long arg)
|
static int exec_simple_method(acpi_handle handle, const char *name, unsigned long arg)
|
||||||
{
|
{
|
||||||
acpi_status status = acpi_execute_simple_method(handle, (char *)name, arg);
|
acpi_status status = acpi_execute_simple_method(handle, (char *)name, arg);
|
||||||
@@ -284,6 +359,89 @@ static int eval_dytc(acpi_handle handle, unsigned long cmd, unsigned long *res)
|
|||||||
return eval_int_with_arg(handle, "DYTC", cmd, res);
|
return eval_int_with_arg(handle, "DYTC", cmd, res);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int eval_vpcr(acpi_handle handle, unsigned long cmd, unsigned long *res)
|
||||||
|
{
|
||||||
|
return eval_int_with_arg(handle, "VPCR", cmd, res);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int eval_vpcw(acpi_handle handle, unsigned long cmd, unsigned long data)
|
||||||
|
{
|
||||||
|
struct acpi_object_list params;
|
||||||
|
union acpi_object in_obj[2];
|
||||||
|
acpi_status status;
|
||||||
|
|
||||||
|
params.count = 2;
|
||||||
|
params.pointer = in_obj;
|
||||||
|
in_obj[0].type = ACPI_TYPE_INTEGER;
|
||||||
|
in_obj[0].integer.value = cmd;
|
||||||
|
in_obj[1].type = ACPI_TYPE_INTEGER;
|
||||||
|
in_obj[1].integer.value = data;
|
||||||
|
|
||||||
|
status = acpi_evaluate_object(handle, "VPCW", ¶ms, NULL);
|
||||||
|
if (ACPI_FAILURE(status))
|
||||||
|
return -EIO;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int read_ec_data(acpi_handle handle, unsigned long cmd, unsigned long *data)
|
||||||
|
{
|
||||||
|
unsigned long end_jiffies, val;
|
||||||
|
int err;
|
||||||
|
|
||||||
|
err = eval_vpcw(handle, 1, cmd);
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
|
end_jiffies = jiffies + msecs_to_jiffies(IDEAPAD_EC_TIMEOUT) + 1;
|
||||||
|
|
||||||
|
while (time_before(jiffies, end_jiffies)) {
|
||||||
|
usleep_range(IDEAPAD_EC_POLL_MIN_US, IDEAPAD_EC_POLL_MAX_US);
|
||||||
|
|
||||||
|
err = eval_vpcr(handle, 1, &val);
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
|
if (val == 0)
|
||||||
|
return eval_vpcr(handle, 0, data);
|
||||||
|
}
|
||||||
|
|
||||||
|
acpi_handle_err(handle, "timeout in %s\n", __func__);
|
||||||
|
|
||||||
|
return -ETIMEDOUT;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int write_ec_cmd(acpi_handle handle, unsigned long cmd, unsigned long data)
|
||||||
|
{
|
||||||
|
unsigned long end_jiffies, val;
|
||||||
|
int err;
|
||||||
|
|
||||||
|
err = eval_vpcw(handle, 0, data);
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
|
err = eval_vpcw(handle, 1, cmd);
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
|
end_jiffies = jiffies + msecs_to_jiffies(IDEAPAD_EC_TIMEOUT) + 1;
|
||||||
|
|
||||||
|
while (time_before(jiffies, end_jiffies)) {
|
||||||
|
usleep_range(IDEAPAD_EC_POLL_MIN_US, IDEAPAD_EC_POLL_MAX_US);
|
||||||
|
|
||||||
|
err = eval_vpcr(handle, 1, &val);
|
||||||
|
if (err)
|
||||||
|
return err;
|
||||||
|
|
||||||
|
if (val == 0)
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
acpi_handle_err(handle, "timeout in %s\n", __func__);
|
||||||
|
|
||||||
|
return -ETIMEDOUT;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* debugfs
|
* debugfs
|
||||||
*/
|
*/
|
||||||
@@ -1501,6 +1659,79 @@ static void ideapad_sync_touchpad_state(struct ideapad_private *priv, bool send_
|
|||||||
priv->r_touchpad_val = value;
|
priv->r_touchpad_val = value;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static const struct dmi_system_id ymc_ec_trigger_quirk_dmi_table[] = {
|
||||||
|
{
|
||||||
|
/* Lenovo Yoga 7 14ARB7 */
|
||||||
|
.matches = {
|
||||||
|
DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
|
||||||
|
DMI_MATCH(DMI_PRODUCT_NAME, "82QF"),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
/* Lenovo Yoga 7 14ACN6 */
|
||||||
|
.matches = {
|
||||||
|
DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
|
||||||
|
DMI_MATCH(DMI_PRODUCT_NAME, "82N7"),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{ }
|
||||||
|
};
|
||||||
|
|
||||||
|
static void ideapad_laptop_trigger_ec(void)
|
||||||
|
{
|
||||||
|
struct ideapad_private *priv;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
guard(mutex)(&ideapad_shared_mutex);
|
||||||
|
|
||||||
|
priv = ideapad_shared;
|
||||||
|
if (!priv)
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (!priv->features.ymc_ec_trigger)
|
||||||
|
return;
|
||||||
|
|
||||||
|
ret = write_ec_cmd(priv->adev->handle, VPCCMD_W_YMC, 1);
|
||||||
|
if (ret)
|
||||||
|
dev_warn(&priv->platform_device->dev, "Could not write YMC: %d\n", ret);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int ideapad_laptop_nb_notify(struct notifier_block *nb,
|
||||||
|
unsigned long action, void *data)
|
||||||
|
{
|
||||||
|
switch (action) {
|
||||||
|
case IDEAPAD_LAPTOP_YMC_EVENT:
|
||||||
|
ideapad_laptop_trigger_ec();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct notifier_block ideapad_laptop_notifier = {
|
||||||
|
.notifier_call = ideapad_laptop_nb_notify,
|
||||||
|
};
|
||||||
|
|
||||||
|
static BLOCKING_NOTIFIER_HEAD(ideapad_laptop_chain_head);
|
||||||
|
|
||||||
|
int ideapad_laptop_register_notifier(struct notifier_block *nb)
|
||||||
|
{
|
||||||
|
return blocking_notifier_chain_register(&ideapad_laptop_chain_head, nb);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_NS_GPL(ideapad_laptop_register_notifier, IDEAPAD_LAPTOP);
|
||||||
|
|
||||||
|
int ideapad_laptop_unregister_notifier(struct notifier_block *nb)
|
||||||
|
{
|
||||||
|
return blocking_notifier_chain_unregister(&ideapad_laptop_chain_head, nb);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_NS_GPL(ideapad_laptop_unregister_notifier, IDEAPAD_LAPTOP);
|
||||||
|
|
||||||
|
void ideapad_laptop_call_notifier(unsigned long action, void *data)
|
||||||
|
{
|
||||||
|
blocking_notifier_call_chain(&ideapad_laptop_chain_head, action, data);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_NS_GPL(ideapad_laptop_call_notifier, IDEAPAD_LAPTOP);
|
||||||
|
|
||||||
static void ideapad_acpi_notify(acpi_handle handle, u32 event, void *data)
|
static void ideapad_acpi_notify(acpi_handle handle, u32 event, void *data)
|
||||||
{
|
{
|
||||||
struct ideapad_private *priv = data;
|
struct ideapad_private *priv = data;
|
||||||
@@ -1637,6 +1868,8 @@ static void ideapad_check_features(struct ideapad_private *priv)
|
|||||||
priv->features.ctrl_ps2_aux_port =
|
priv->features.ctrl_ps2_aux_port =
|
||||||
ctrl_ps2_aux_port || dmi_check_system(ctrl_ps2_aux_port_list);
|
ctrl_ps2_aux_port || dmi_check_system(ctrl_ps2_aux_port_list);
|
||||||
priv->features.touchpad_ctrl_via_ec = touchpad_ctrl_via_ec;
|
priv->features.touchpad_ctrl_via_ec = touchpad_ctrl_via_ec;
|
||||||
|
priv->features.ymc_ec_trigger =
|
||||||
|
ymc_ec_trigger || dmi_check_system(ymc_ec_trigger_quirk_dmi_table);
|
||||||
|
|
||||||
if (!read_ec_data(handle, VPCCMD_R_FAN, &val))
|
if (!read_ec_data(handle, VPCCMD_R_FAN, &val))
|
||||||
priv->features.fan_mode = true;
|
priv->features.fan_mode = true;
|
||||||
@@ -1872,6 +2105,8 @@ static int ideapad_acpi_add(struct platform_device *pdev)
|
|||||||
if (err)
|
if (err)
|
||||||
goto shared_init_failed;
|
goto shared_init_failed;
|
||||||
|
|
||||||
|
ideapad_laptop_register_notifier(&ideapad_laptop_notifier);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
shared_init_failed:
|
shared_init_failed:
|
||||||
@@ -1903,6 +2138,8 @@ static void ideapad_acpi_remove(struct platform_device *pdev)
|
|||||||
struct ideapad_private *priv = dev_get_drvdata(&pdev->dev);
|
struct ideapad_private *priv = dev_get_drvdata(&pdev->dev);
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
|
ideapad_laptop_unregister_notifier(&ideapad_laptop_notifier);
|
||||||
|
|
||||||
ideapad_shared_exit(priv);
|
ideapad_shared_exit(priv);
|
||||||
|
|
||||||
acpi_remove_notify_handler(priv->adev->handle,
|
acpi_remove_notify_handler(priv->adev->handle,
|
||||||
|
@@ -9,144 +9,14 @@
|
|||||||
#ifndef _IDEAPAD_LAPTOP_H_
|
#ifndef _IDEAPAD_LAPTOP_H_
|
||||||
#define _IDEAPAD_LAPTOP_H_
|
#define _IDEAPAD_LAPTOP_H_
|
||||||
|
|
||||||
#include <linux/acpi.h>
|
#include <linux/notifier.h>
|
||||||
#include <linux/jiffies.h>
|
|
||||||
#include <linux/errno.h>
|
|
||||||
|
|
||||||
enum {
|
enum ideapad_laptop_notifier_actions {
|
||||||
VPCCMD_R_VPC1 = 0x10,
|
IDEAPAD_LAPTOP_YMC_EVENT,
|
||||||
VPCCMD_R_BL_MAX,
|
|
||||||
VPCCMD_R_BL,
|
|
||||||
VPCCMD_W_BL,
|
|
||||||
VPCCMD_R_WIFI,
|
|
||||||
VPCCMD_W_WIFI,
|
|
||||||
VPCCMD_R_BT,
|
|
||||||
VPCCMD_W_BT,
|
|
||||||
VPCCMD_R_BL_POWER,
|
|
||||||
VPCCMD_R_NOVO,
|
|
||||||
VPCCMD_R_VPC2,
|
|
||||||
VPCCMD_R_TOUCHPAD,
|
|
||||||
VPCCMD_W_TOUCHPAD,
|
|
||||||
VPCCMD_R_CAMERA,
|
|
||||||
VPCCMD_W_CAMERA,
|
|
||||||
VPCCMD_R_3G,
|
|
||||||
VPCCMD_W_3G,
|
|
||||||
VPCCMD_R_ODD, /* 0x21 */
|
|
||||||
VPCCMD_W_FAN,
|
|
||||||
VPCCMD_R_RF,
|
|
||||||
VPCCMD_W_RF,
|
|
||||||
VPCCMD_W_YMC = 0x2A,
|
|
||||||
VPCCMD_R_FAN = 0x2B,
|
|
||||||
VPCCMD_R_SPECIAL_BUTTONS = 0x31,
|
|
||||||
VPCCMD_W_BL_POWER = 0x33,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline int eval_int_with_arg(acpi_handle handle, const char *name, unsigned long arg, unsigned long *res)
|
int ideapad_laptop_register_notifier(struct notifier_block *nb);
|
||||||
{
|
int ideapad_laptop_unregister_notifier(struct notifier_block *nb);
|
||||||
struct acpi_object_list params;
|
void ideapad_laptop_call_notifier(unsigned long action, void *data);
|
||||||
unsigned long long result;
|
|
||||||
union acpi_object in_obj;
|
|
||||||
acpi_status status;
|
|
||||||
|
|
||||||
params.count = 1;
|
|
||||||
params.pointer = &in_obj;
|
|
||||||
in_obj.type = ACPI_TYPE_INTEGER;
|
|
||||||
in_obj.integer.value = arg;
|
|
||||||
|
|
||||||
status = acpi_evaluate_integer(handle, (char *)name, ¶ms, &result);
|
|
||||||
if (ACPI_FAILURE(status))
|
|
||||||
return -EIO;
|
|
||||||
|
|
||||||
if (res)
|
|
||||||
*res = result;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int eval_vpcr(acpi_handle handle, unsigned long cmd, unsigned long *res)
|
|
||||||
{
|
|
||||||
return eval_int_with_arg(handle, "VPCR", cmd, res);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int eval_vpcw(acpi_handle handle, unsigned long cmd, unsigned long data)
|
|
||||||
{
|
|
||||||
struct acpi_object_list params;
|
|
||||||
union acpi_object in_obj[2];
|
|
||||||
acpi_status status;
|
|
||||||
|
|
||||||
params.count = 2;
|
|
||||||
params.pointer = in_obj;
|
|
||||||
in_obj[0].type = ACPI_TYPE_INTEGER;
|
|
||||||
in_obj[0].integer.value = cmd;
|
|
||||||
in_obj[1].type = ACPI_TYPE_INTEGER;
|
|
||||||
in_obj[1].integer.value = data;
|
|
||||||
|
|
||||||
status = acpi_evaluate_object(handle, "VPCW", ¶ms, NULL);
|
|
||||||
if (ACPI_FAILURE(status))
|
|
||||||
return -EIO;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
#define IDEAPAD_EC_TIMEOUT 200 /* in ms */
|
|
||||||
|
|
||||||
static inline int read_ec_data(acpi_handle handle, unsigned long cmd, unsigned long *data)
|
|
||||||
{
|
|
||||||
unsigned long end_jiffies, val;
|
|
||||||
int err;
|
|
||||||
|
|
||||||
err = eval_vpcw(handle, 1, cmd);
|
|
||||||
if (err)
|
|
||||||
return err;
|
|
||||||
|
|
||||||
end_jiffies = jiffies + msecs_to_jiffies(IDEAPAD_EC_TIMEOUT) + 1;
|
|
||||||
|
|
||||||
while (time_before(jiffies, end_jiffies)) {
|
|
||||||
schedule();
|
|
||||||
|
|
||||||
err = eval_vpcr(handle, 1, &val);
|
|
||||||
if (err)
|
|
||||||
return err;
|
|
||||||
|
|
||||||
if (val == 0)
|
|
||||||
return eval_vpcr(handle, 0, data);
|
|
||||||
}
|
|
||||||
|
|
||||||
acpi_handle_err(handle, "timeout in %s\n", __func__);
|
|
||||||
|
|
||||||
return -ETIMEDOUT;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int write_ec_cmd(acpi_handle handle, unsigned long cmd, unsigned long data)
|
|
||||||
{
|
|
||||||
unsigned long end_jiffies, val;
|
|
||||||
int err;
|
|
||||||
|
|
||||||
err = eval_vpcw(handle, 0, data);
|
|
||||||
if (err)
|
|
||||||
return err;
|
|
||||||
|
|
||||||
err = eval_vpcw(handle, 1, cmd);
|
|
||||||
if (err)
|
|
||||||
return err;
|
|
||||||
|
|
||||||
end_jiffies = jiffies + msecs_to_jiffies(IDEAPAD_EC_TIMEOUT) + 1;
|
|
||||||
|
|
||||||
while (time_before(jiffies, end_jiffies)) {
|
|
||||||
schedule();
|
|
||||||
|
|
||||||
err = eval_vpcr(handle, 1, &val);
|
|
||||||
if (err)
|
|
||||||
return err;
|
|
||||||
|
|
||||||
if (val == 0)
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
acpi_handle_err(handle, "timeout in %s\n", __func__);
|
|
||||||
|
|
||||||
return -ETIMEDOUT;
|
|
||||||
}
|
|
||||||
|
|
||||||
#undef IDEAPAD_EC_TIMEOUT
|
|
||||||
#endif /* !_IDEAPAD_LAPTOP_H_ */
|
#endif /* !_IDEAPAD_LAPTOP_H_ */
|
||||||
|
@@ -20,32 +20,10 @@
|
|||||||
#define LENOVO_YMC_QUERY_INSTANCE 0
|
#define LENOVO_YMC_QUERY_INSTANCE 0
|
||||||
#define LENOVO_YMC_QUERY_METHOD 0x01
|
#define LENOVO_YMC_QUERY_METHOD 0x01
|
||||||
|
|
||||||
static bool ec_trigger __read_mostly;
|
|
||||||
module_param(ec_trigger, bool, 0444);
|
|
||||||
MODULE_PARM_DESC(ec_trigger, "Enable EC triggering work-around to force emitting tablet mode events");
|
|
||||||
|
|
||||||
static bool force;
|
static bool force;
|
||||||
module_param(force, bool, 0444);
|
module_param(force, bool, 0444);
|
||||||
MODULE_PARM_DESC(force, "Force loading on boards without a convertible DMI chassis-type");
|
MODULE_PARM_DESC(force, "Force loading on boards without a convertible DMI chassis-type");
|
||||||
|
|
||||||
static const struct dmi_system_id ec_trigger_quirk_dmi_table[] = {
|
|
||||||
{
|
|
||||||
/* Lenovo Yoga 7 14ARB7 */
|
|
||||||
.matches = {
|
|
||||||
DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
|
|
||||||
DMI_MATCH(DMI_PRODUCT_NAME, "82QF"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
/* Lenovo Yoga 7 14ACN6 */
|
|
||||||
.matches = {
|
|
||||||
DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
|
|
||||||
DMI_MATCH(DMI_PRODUCT_NAME, "82N7"),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{ }
|
|
||||||
};
|
|
||||||
|
|
||||||
static const struct dmi_system_id allowed_chasis_types_dmi_table[] = {
|
static const struct dmi_system_id allowed_chasis_types_dmi_table[] = {
|
||||||
{
|
{
|
||||||
.matches = {
|
.matches = {
|
||||||
@@ -62,21 +40,8 @@ static const struct dmi_system_id allowed_chasis_types_dmi_table[] = {
|
|||||||
|
|
||||||
struct lenovo_ymc_private {
|
struct lenovo_ymc_private {
|
||||||
struct input_dev *input_dev;
|
struct input_dev *input_dev;
|
||||||
struct acpi_device *ec_acpi_dev;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static void lenovo_ymc_trigger_ec(struct wmi_device *wdev, struct lenovo_ymc_private *priv)
|
|
||||||
{
|
|
||||||
int err;
|
|
||||||
|
|
||||||
if (!priv->ec_acpi_dev)
|
|
||||||
return;
|
|
||||||
|
|
||||||
err = write_ec_cmd(priv->ec_acpi_dev->handle, VPCCMD_W_YMC, 1);
|
|
||||||
if (err)
|
|
||||||
dev_warn(&wdev->dev, "Could not write YMC: %d\n", err);
|
|
||||||
}
|
|
||||||
|
|
||||||
static const struct key_entry lenovo_ymc_keymap[] = {
|
static const struct key_entry lenovo_ymc_keymap[] = {
|
||||||
/* Ignore the uninitialized state */
|
/* Ignore the uninitialized state */
|
||||||
{ KE_IGNORE, 0x00 },
|
{ KE_IGNORE, 0x00 },
|
||||||
@@ -127,11 +92,9 @@ static void lenovo_ymc_notify(struct wmi_device *wdev, union acpi_object *data)
|
|||||||
|
|
||||||
free_obj:
|
free_obj:
|
||||||
kfree(obj);
|
kfree(obj);
|
||||||
lenovo_ymc_trigger_ec(wdev, priv);
|
ideapad_laptop_call_notifier(IDEAPAD_LAPTOP_YMC_EVENT, &code);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void acpi_dev_put_helper(void *p) { acpi_dev_put(p); }
|
|
||||||
|
|
||||||
static int lenovo_ymc_probe(struct wmi_device *wdev, const void *ctx)
|
static int lenovo_ymc_probe(struct wmi_device *wdev, const void *ctx)
|
||||||
{
|
{
|
||||||
struct lenovo_ymc_private *priv;
|
struct lenovo_ymc_private *priv;
|
||||||
@@ -145,29 +108,10 @@ static int lenovo_ymc_probe(struct wmi_device *wdev, const void *ctx)
|
|||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
}
|
}
|
||||||
|
|
||||||
ec_trigger |= dmi_check_system(ec_trigger_quirk_dmi_table);
|
|
||||||
|
|
||||||
priv = devm_kzalloc(&wdev->dev, sizeof(*priv), GFP_KERNEL);
|
priv = devm_kzalloc(&wdev->dev, sizeof(*priv), GFP_KERNEL);
|
||||||
if (!priv)
|
if (!priv)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
if (ec_trigger) {
|
|
||||||
pr_debug("Lenovo YMC enable EC triggering.\n");
|
|
||||||
priv->ec_acpi_dev = acpi_dev_get_first_match_dev("VPC2004", NULL, -1);
|
|
||||||
|
|
||||||
if (!priv->ec_acpi_dev) {
|
|
||||||
dev_err(&wdev->dev, "Could not find EC ACPI device.\n");
|
|
||||||
return -ENODEV;
|
|
||||||
}
|
|
||||||
err = devm_add_action_or_reset(&wdev->dev,
|
|
||||||
acpi_dev_put_helper, priv->ec_acpi_dev);
|
|
||||||
if (err) {
|
|
||||||
dev_err(&wdev->dev,
|
|
||||||
"Could not clean up EC ACPI device: %d\n", err);
|
|
||||||
return err;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
input_dev = devm_input_allocate_device(&wdev->dev);
|
input_dev = devm_input_allocate_device(&wdev->dev);
|
||||||
if (!input_dev)
|
if (!input_dev)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
@@ -194,7 +138,6 @@ static int lenovo_ymc_probe(struct wmi_device *wdev, const void *ctx)
|
|||||||
dev_set_drvdata(&wdev->dev, priv);
|
dev_set_drvdata(&wdev->dev, priv);
|
||||||
|
|
||||||
/* Report the state for the first time on probe */
|
/* Report the state for the first time on probe */
|
||||||
lenovo_ymc_trigger_ec(wdev, priv);
|
|
||||||
lenovo_ymc_notify(wdev, NULL);
|
lenovo_ymc_notify(wdev, NULL);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@@ -219,3 +162,4 @@ module_wmi_driver(lenovo_ymc_driver);
|
|||||||
MODULE_AUTHOR("Gergo Koteles <soyer@irl.hu>");
|
MODULE_AUTHOR("Gergo Koteles <soyer@irl.hu>");
|
||||||
MODULE_DESCRIPTION("Lenovo Yoga Mode Control driver");
|
MODULE_DESCRIPTION("Lenovo Yoga Mode Control driver");
|
||||||
MODULE_LICENSE("GPL");
|
MODULE_LICENSE("GPL");
|
||||||
|
MODULE_IMPORT_NS(IDEAPAD_LAPTOP);
|
||||||
|
@@ -1333,7 +1333,7 @@ static void *_copy_apqns_from_user(void __user *uapqns, size_t nr_apqns)
|
|||||||
if (!uapqns || nr_apqns == 0)
|
if (!uapqns || nr_apqns == 0)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
return memdup_user(uapqns, nr_apqns * sizeof(struct pkey_apqn));
|
return memdup_array_user(uapqns, nr_apqns, sizeof(struct pkey_apqn));
|
||||||
}
|
}
|
||||||
|
|
||||||
static long pkey_unlocked_ioctl(struct file *filp, unsigned int cmd,
|
static long pkey_unlocked_ioctl(struct file *filp, unsigned int cmd,
|
||||||
|
@@ -5908,7 +5908,11 @@ megasas_set_high_iops_queue_affinity_and_hint(struct megasas_instance *instance)
|
|||||||
const struct cpumask *mask;
|
const struct cpumask *mask;
|
||||||
|
|
||||||
if (instance->perf_mode == MR_BALANCED_PERF_MODE) {
|
if (instance->perf_mode == MR_BALANCED_PERF_MODE) {
|
||||||
mask = cpumask_of_node(dev_to_node(&instance->pdev->dev));
|
int nid = dev_to_node(&instance->pdev->dev);
|
||||||
|
|
||||||
|
if (nid == NUMA_NO_NODE)
|
||||||
|
nid = 0;
|
||||||
|
mask = cpumask_of_node(nid);
|
||||||
|
|
||||||
for (i = 0; i < instance->low_latency_index_start; i++) {
|
for (i = 0; i < instance->low_latency_index_start; i++) {
|
||||||
irq = pci_irq_vector(instance->pdev, i);
|
irq = pci_irq_vector(instance->pdev, i);
|
||||||
|
@@ -1868,6 +1868,13 @@ static int cqspi_probe(struct platform_device *pdev)
|
|||||||
goto probe_setup_failed;
|
goto probe_setup_failed;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pm_runtime_enable(dev);
|
||||||
|
|
||||||
|
if (cqspi->rx_chan) {
|
||||||
|
dma_release_channel(cqspi->rx_chan);
|
||||||
|
goto probe_setup_failed;
|
||||||
|
}
|
||||||
|
|
||||||
ret = spi_register_controller(host);
|
ret = spi_register_controller(host);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
dev_err(&pdev->dev, "failed to register SPI ctlr %d\n", ret);
|
dev_err(&pdev->dev, "failed to register SPI ctlr %d\n", ret);
|
||||||
@@ -1877,6 +1884,7 @@ static int cqspi_probe(struct platform_device *pdev)
|
|||||||
return 0;
|
return 0;
|
||||||
probe_setup_failed:
|
probe_setup_failed:
|
||||||
cqspi_controller_enable(cqspi, 0);
|
cqspi_controller_enable(cqspi, 0);
|
||||||
|
pm_runtime_disable(dev);
|
||||||
probe_reset_failed:
|
probe_reset_failed:
|
||||||
if (cqspi->is_jh7110)
|
if (cqspi->is_jh7110)
|
||||||
cqspi_jh7110_disable_clk(pdev, cqspi);
|
cqspi_jh7110_disable_clk(pdev, cqspi);
|
||||||
@@ -1898,7 +1906,8 @@ static void cqspi_remove(struct platform_device *pdev)
|
|||||||
if (cqspi->rx_chan)
|
if (cqspi->rx_chan)
|
||||||
dma_release_channel(cqspi->rx_chan);
|
dma_release_channel(cqspi->rx_chan);
|
||||||
|
|
||||||
clk_disable_unprepare(cqspi->clk);
|
if (pm_runtime_get_sync(&pdev->dev) >= 0)
|
||||||
|
clk_disable(cqspi->clk);
|
||||||
|
|
||||||
if (cqspi->is_jh7110)
|
if (cqspi->is_jh7110)
|
||||||
cqspi_jh7110_disable_clk(pdev, cqspi);
|
cqspi_jh7110_disable_clk(pdev, cqspi);
|
||||||
|
@@ -869,29 +869,21 @@ static signed int aes_cipher(u8 *key, uint hdrlen,
|
|||||||
num_blocks, payload_index;
|
num_blocks, payload_index;
|
||||||
|
|
||||||
u8 pn_vector[6];
|
u8 pn_vector[6];
|
||||||
u8 mic_iv[16];
|
u8 mic_iv[16] = {};
|
||||||
u8 mic_header1[16];
|
u8 mic_header1[16] = {};
|
||||||
u8 mic_header2[16];
|
u8 mic_header2[16] = {};
|
||||||
u8 ctr_preload[16];
|
u8 ctr_preload[16] = {};
|
||||||
|
|
||||||
/* Intermediate Buffers */
|
/* Intermediate Buffers */
|
||||||
u8 chain_buffer[16];
|
u8 chain_buffer[16] = {};
|
||||||
u8 aes_out[16];
|
u8 aes_out[16] = {};
|
||||||
u8 padded_buffer[16];
|
u8 padded_buffer[16] = {};
|
||||||
u8 mic[8];
|
u8 mic[8];
|
||||||
uint frtype = GetFrameType(pframe);
|
uint frtype = GetFrameType(pframe);
|
||||||
uint frsubtype = GetFrameSubType(pframe);
|
uint frsubtype = GetFrameSubType(pframe);
|
||||||
|
|
||||||
frsubtype = frsubtype>>4;
|
frsubtype = frsubtype>>4;
|
||||||
|
|
||||||
memset((void *)mic_iv, 0, 16);
|
|
||||||
memset((void *)mic_header1, 0, 16);
|
|
||||||
memset((void *)mic_header2, 0, 16);
|
|
||||||
memset((void *)ctr_preload, 0, 16);
|
|
||||||
memset((void *)chain_buffer, 0, 16);
|
|
||||||
memset((void *)aes_out, 0, 16);
|
|
||||||
memset((void *)padded_buffer, 0, 16);
|
|
||||||
|
|
||||||
if ((hdrlen == WLAN_HDR_A3_LEN) || (hdrlen == WLAN_HDR_A3_QOS_LEN))
|
if ((hdrlen == WLAN_HDR_A3_LEN) || (hdrlen == WLAN_HDR_A3_QOS_LEN))
|
||||||
a4_exists = 0;
|
a4_exists = 0;
|
||||||
else
|
else
|
||||||
@@ -1081,15 +1073,15 @@ static signed int aes_decipher(u8 *key, uint hdrlen,
|
|||||||
num_blocks, payload_index;
|
num_blocks, payload_index;
|
||||||
signed int res = _SUCCESS;
|
signed int res = _SUCCESS;
|
||||||
u8 pn_vector[6];
|
u8 pn_vector[6];
|
||||||
u8 mic_iv[16];
|
u8 mic_iv[16] = {};
|
||||||
u8 mic_header1[16];
|
u8 mic_header1[16] = {};
|
||||||
u8 mic_header2[16];
|
u8 mic_header2[16] = {};
|
||||||
u8 ctr_preload[16];
|
u8 ctr_preload[16] = {};
|
||||||
|
|
||||||
/* Intermediate Buffers */
|
/* Intermediate Buffers */
|
||||||
u8 chain_buffer[16];
|
u8 chain_buffer[16] = {};
|
||||||
u8 aes_out[16];
|
u8 aes_out[16] = {};
|
||||||
u8 padded_buffer[16];
|
u8 padded_buffer[16] = {};
|
||||||
u8 mic[8];
|
u8 mic[8];
|
||||||
|
|
||||||
uint frtype = GetFrameType(pframe);
|
uint frtype = GetFrameType(pframe);
|
||||||
@@ -1097,14 +1089,6 @@ static signed int aes_decipher(u8 *key, uint hdrlen,
|
|||||||
|
|
||||||
frsubtype = frsubtype>>4;
|
frsubtype = frsubtype>>4;
|
||||||
|
|
||||||
memset((void *)mic_iv, 0, 16);
|
|
||||||
memset((void *)mic_header1, 0, 16);
|
|
||||||
memset((void *)mic_header2, 0, 16);
|
|
||||||
memset((void *)ctr_preload, 0, 16);
|
|
||||||
memset((void *)chain_buffer, 0, 16);
|
|
||||||
memset((void *)aes_out, 0, 16);
|
|
||||||
memset((void *)padded_buffer, 0, 16);
|
|
||||||
|
|
||||||
/* start to decrypt the payload */
|
/* start to decrypt the payload */
|
||||||
|
|
||||||
num_blocks = (plen-8) / 16; /* plen including LLC, payload_length and mic) */
|
num_blocks = (plen-8) / 16; /* plen including LLC, payload_length and mic) */
|
||||||
|
@@ -234,6 +234,7 @@ struct imx_port {
|
|||||||
enum imx_tx_state tx_state;
|
enum imx_tx_state tx_state;
|
||||||
struct hrtimer trigger_start_tx;
|
struct hrtimer trigger_start_tx;
|
||||||
struct hrtimer trigger_stop_tx;
|
struct hrtimer trigger_stop_tx;
|
||||||
|
unsigned int rxtl;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct imx_port_ucrs {
|
struct imx_port_ucrs {
|
||||||
@@ -1337,6 +1338,7 @@ static void imx_uart_clear_rx_errors(struct imx_port *sport)
|
|||||||
|
|
||||||
#define TXTL_DEFAULT 8
|
#define TXTL_DEFAULT 8
|
||||||
#define RXTL_DEFAULT 8 /* 8 characters or aging timer */
|
#define RXTL_DEFAULT 8 /* 8 characters or aging timer */
|
||||||
|
#define RXTL_CONSOLE_DEFAULT 1
|
||||||
#define TXTL_DMA 8 /* DMA burst setting */
|
#define TXTL_DMA 8 /* DMA burst setting */
|
||||||
#define RXTL_DMA 9 /* DMA burst setting */
|
#define RXTL_DMA 9 /* DMA burst setting */
|
||||||
|
|
||||||
@@ -1449,7 +1451,7 @@ static void imx_uart_disable_dma(struct imx_port *sport)
|
|||||||
ucr1 &= ~(UCR1_RXDMAEN | UCR1_TXDMAEN | UCR1_ATDMAEN);
|
ucr1 &= ~(UCR1_RXDMAEN | UCR1_TXDMAEN | UCR1_ATDMAEN);
|
||||||
imx_uart_writel(sport, ucr1, UCR1);
|
imx_uart_writel(sport, ucr1, UCR1);
|
||||||
|
|
||||||
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, RXTL_DEFAULT);
|
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, sport->rxtl);
|
||||||
|
|
||||||
sport->dma_is_enabled = 0;
|
sport->dma_is_enabled = 0;
|
||||||
}
|
}
|
||||||
@@ -1474,7 +1476,12 @@ static int imx_uart_startup(struct uart_port *port)
|
|||||||
return retval;
|
return retval;
|
||||||
}
|
}
|
||||||
|
|
||||||
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, RXTL_DEFAULT);
|
if (uart_console(&sport->port))
|
||||||
|
sport->rxtl = RXTL_CONSOLE_DEFAULT;
|
||||||
|
else
|
||||||
|
sport->rxtl = RXTL_DEFAULT;
|
||||||
|
|
||||||
|
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, sport->rxtl);
|
||||||
|
|
||||||
/* disable the DREN bit (Data Ready interrupt enable) before
|
/* disable the DREN bit (Data Ready interrupt enable) before
|
||||||
* requesting IRQs
|
* requesting IRQs
|
||||||
@@ -1887,7 +1894,7 @@ static int imx_uart_poll_init(struct uart_port *port)
|
|||||||
if (retval)
|
if (retval)
|
||||||
clk_disable_unprepare(sport->clk_ipg);
|
clk_disable_unprepare(sport->clk_ipg);
|
||||||
|
|
||||||
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, RXTL_DEFAULT);
|
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, sport->rxtl);
|
||||||
|
|
||||||
spin_lock_irqsave(&sport->port.lock, flags);
|
spin_lock_irqsave(&sport->port.lock, flags);
|
||||||
|
|
||||||
@@ -1979,7 +1986,7 @@ static int imx_uart_rs485_config(struct uart_port *port, struct ktermios *termio
|
|||||||
/* If the receiver trigger is 0, set it to a default value */
|
/* If the receiver trigger is 0, set it to a default value */
|
||||||
ufcr = imx_uart_readl(sport, UFCR);
|
ufcr = imx_uart_readl(sport, UFCR);
|
||||||
if ((ufcr & UFCR_RXTL_MASK) == 0)
|
if ((ufcr & UFCR_RXTL_MASK) == 0)
|
||||||
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, RXTL_DEFAULT);
|
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, sport->rxtl);
|
||||||
imx_uart_start_rx(port);
|
imx_uart_start_rx(port);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2164,7 +2171,7 @@ imx_uart_console_setup(struct console *co, char *options)
|
|||||||
else
|
else
|
||||||
imx_uart_console_get_options(sport, &baud, &parity, &bits);
|
imx_uart_console_get_options(sport, &baud, &parity, &bits);
|
||||||
|
|
||||||
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, RXTL_DEFAULT);
|
imx_uart_setup_ufcr(sport, TXTL_DEFAULT, sport->rxtl);
|
||||||
|
|
||||||
retval = uart_set_options(&sport->port, co, baud, parity, bits, flow);
|
retval = uart_set_options(&sport->port, co, baud, parity, bits, flow);
|
||||||
|
|
||||||
|
@@ -872,16 +872,6 @@ of_err:
|
|||||||
pm_runtime_set_active(&pdev->dev);
|
pm_runtime_set_active(&pdev->dev);
|
||||||
pm_runtime_enable(&pdev->dev);
|
pm_runtime_enable(&pdev->dev);
|
||||||
|
|
||||||
if (!ulite_uart_driver.state) {
|
|
||||||
dev_dbg(&pdev->dev, "uartlite: calling uart_register_driver()\n");
|
|
||||||
ret = uart_register_driver(&ulite_uart_driver);
|
|
||||||
if (ret < 0) {
|
|
||||||
dev_err(&pdev->dev, "Failed to register driver\n");
|
|
||||||
clk_disable_unprepare(pdata->clk);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
ret = ulite_assign(&pdev->dev, id, res->start, irq, pdata);
|
ret = ulite_assign(&pdev->dev, id, res->start, irq, pdata);
|
||||||
|
|
||||||
pm_runtime_mark_last_busy(&pdev->dev);
|
pm_runtime_mark_last_busy(&pdev->dev);
|
||||||
@@ -922,15 +912,24 @@ static struct platform_driver ulite_platform_driver = {
|
|||||||
|
|
||||||
static int __init ulite_init(void)
|
static int __init ulite_init(void)
|
||||||
{
|
{
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
pr_debug("uartlite: calling uart_register_driver()\n");
|
||||||
|
ret = uart_register_driver(&ulite_uart_driver);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
pr_debug("uartlite: calling platform_driver_register()\n");
|
pr_debug("uartlite: calling platform_driver_register()\n");
|
||||||
return platform_driver_register(&ulite_platform_driver);
|
ret = platform_driver_register(&ulite_platform_driver);
|
||||||
|
if (ret)
|
||||||
|
uart_unregister_driver(&ulite_uart_driver);
|
||||||
|
|
||||||
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __exit ulite_exit(void)
|
static void __exit ulite_exit(void)
|
||||||
{
|
{
|
||||||
platform_driver_unregister(&ulite_platform_driver);
|
platform_driver_unregister(&ulite_platform_driver);
|
||||||
if (ulite_uart_driver.state)
|
|
||||||
uart_unregister_driver(&ulite_uart_driver);
|
uart_unregister_driver(&ulite_uart_driver);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -962,7 +962,7 @@ void redraw_screen(struct vc_data *vc, int is_switch)
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (redraw) {
|
if (redraw) {
|
||||||
int update;
|
bool update;
|
||||||
int old_was_color = vc->vc_can_do_color;
|
int old_was_color = vc->vc_can_do_color;
|
||||||
|
|
||||||
set_origin(vc);
|
set_origin(vc);
|
||||||
@@ -999,7 +999,7 @@ int vc_cons_allocated(unsigned int i)
|
|||||||
return (i < MAX_NR_CONSOLES && vc_cons[i].d);
|
return (i < MAX_NR_CONSOLES && vc_cons[i].d);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void visual_init(struct vc_data *vc, int num, int init)
|
static void visual_init(struct vc_data *vc, int num, bool init)
|
||||||
{
|
{
|
||||||
/* ++Geert: vc->vc_sw->con_init determines console size */
|
/* ++Geert: vc->vc_sw->con_init determines console size */
|
||||||
if (vc->vc_sw)
|
if (vc->vc_sw)
|
||||||
@@ -1083,7 +1083,7 @@ int vc_allocate(unsigned int currcons) /* return 0 on success */
|
|||||||
vc->port.ops = &vc_port_ops;
|
vc->port.ops = &vc_port_ops;
|
||||||
INIT_WORK(&vc_cons[currcons].SAK_work, vc_SAK);
|
INIT_WORK(&vc_cons[currcons].SAK_work, vc_SAK);
|
||||||
|
|
||||||
visual_init(vc, currcons, 1);
|
visual_init(vc, currcons, true);
|
||||||
|
|
||||||
if (!*vc->uni_pagedict_loc)
|
if (!*vc->uni_pagedict_loc)
|
||||||
con_set_default_unimap(vc);
|
con_set_default_unimap(vc);
|
||||||
@@ -1582,7 +1582,7 @@ static void csi_X(struct vc_data *vc, unsigned int vpar)
|
|||||||
vc_uniscr_clear_line(vc, vc->state.x, count);
|
vc_uniscr_clear_line(vc, vc->state.x, count);
|
||||||
scr_memsetw((unsigned short *)vc->vc_pos, vc->vc_video_erase_char, 2 * count);
|
scr_memsetw((unsigned short *)vc->vc_pos, vc->vc_video_erase_char, 2 * count);
|
||||||
if (con_should_update(vc))
|
if (con_should_update(vc))
|
||||||
vc->vc_sw->con_clear(vc, vc->state.y, vc->state.x, 1, count);
|
vc->vc_sw->con_clear(vc, vc->state.y, vc->state.x, count);
|
||||||
vc->vc_need_wrap = 0;
|
vc->vc_need_wrap = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -3474,7 +3474,7 @@ static int __init con_init(void)
|
|||||||
vc_cons[currcons].d = vc = kzalloc(sizeof(struct vc_data), GFP_NOWAIT);
|
vc_cons[currcons].d = vc = kzalloc(sizeof(struct vc_data), GFP_NOWAIT);
|
||||||
INIT_WORK(&vc_cons[currcons].SAK_work, vc_SAK);
|
INIT_WORK(&vc_cons[currcons].SAK_work, vc_SAK);
|
||||||
tty_port_init(&vc->port);
|
tty_port_init(&vc->port);
|
||||||
visual_init(vc, currcons, 1);
|
visual_init(vc, currcons, true);
|
||||||
/* Assuming vc->vc_{cols,rows,screenbuf_size} are sane here. */
|
/* Assuming vc->vc_{cols,rows,screenbuf_size} are sane here. */
|
||||||
vc->vc_screenbuf = kzalloc(vc->vc_screenbuf_size, GFP_NOWAIT);
|
vc->vc_screenbuf = kzalloc(vc->vc_screenbuf_size, GFP_NOWAIT);
|
||||||
vc_init(vc, currcons || !vc->vc_sw->con_save_screen);
|
vc_init(vc, currcons || !vc->vc_sw->con_save_screen);
|
||||||
@@ -3642,7 +3642,7 @@ static int do_bind_con_driver(const struct consw *csw, int first, int last,
|
|||||||
old_was_color = vc->vc_can_do_color;
|
old_was_color = vc->vc_can_do_color;
|
||||||
vc->vc_sw->con_deinit(vc);
|
vc->vc_sw->con_deinit(vc);
|
||||||
vc->vc_origin = (unsigned long)vc->vc_screenbuf;
|
vc->vc_origin = (unsigned long)vc->vc_screenbuf;
|
||||||
visual_init(vc, i, 0);
|
visual_init(vc, i, false);
|
||||||
set_origin(vc);
|
set_origin(vc);
|
||||||
update_attr(vc);
|
update_attr(vc);
|
||||||
|
|
||||||
|
@@ -1319,6 +1319,7 @@ static int ufshcd_clock_scaling_prepare(struct ufs_hba *hba, u64 timeout_us)
|
|||||||
* make sure that there are no outstanding requests when
|
* make sure that there are no outstanding requests when
|
||||||
* clock scaling is in progress
|
* clock scaling is in progress
|
||||||
*/
|
*/
|
||||||
|
mutex_lock(&hba->host->scan_mutex);
|
||||||
blk_mq_quiesce_tagset(&hba->host->tag_set);
|
blk_mq_quiesce_tagset(&hba->host->tag_set);
|
||||||
mutex_lock(&hba->wb_mutex);
|
mutex_lock(&hba->wb_mutex);
|
||||||
down_write(&hba->clk_scaling_lock);
|
down_write(&hba->clk_scaling_lock);
|
||||||
@@ -1329,6 +1330,7 @@ static int ufshcd_clock_scaling_prepare(struct ufs_hba *hba, u64 timeout_us)
|
|||||||
up_write(&hba->clk_scaling_lock);
|
up_write(&hba->clk_scaling_lock);
|
||||||
mutex_unlock(&hba->wb_mutex);
|
mutex_unlock(&hba->wb_mutex);
|
||||||
blk_mq_unquiesce_tagset(&hba->host->tag_set);
|
blk_mq_unquiesce_tagset(&hba->host->tag_set);
|
||||||
|
mutex_unlock(&hba->host->scan_mutex);
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1350,6 +1352,7 @@ static void ufshcd_clock_scaling_unprepare(struct ufs_hba *hba, int err, bool sc
|
|||||||
mutex_unlock(&hba->wb_mutex);
|
mutex_unlock(&hba->wb_mutex);
|
||||||
|
|
||||||
blk_mq_unquiesce_tagset(&hba->host->tag_set);
|
blk_mq_unquiesce_tagset(&hba->host->tag_set);
|
||||||
|
mutex_unlock(&hba->host->scan_mutex);
|
||||||
ufshcd_release(hba);
|
ufshcd_release(hba);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -249,6 +249,7 @@ hv_uio_probe(struct hv_device *dev,
|
|||||||
struct hv_uio_private_data *pdata;
|
struct hv_uio_private_data *pdata;
|
||||||
void *ring_buffer;
|
void *ring_buffer;
|
||||||
int ret;
|
int ret;
|
||||||
|
size_t ring_size = hv_dev_ring_size(channel);
|
||||||
|
|
||||||
/* Communicating with host has to be via shared memory not hypercall */
|
/* Communicating with host has to be via shared memory not hypercall */
|
||||||
if (!channel->offermsg.monitor_allocated) {
|
if (!channel->offermsg.monitor_allocated) {
|
||||||
@@ -256,12 +257,17 @@ hv_uio_probe(struct hv_device *dev,
|
|||||||
return -ENOTSUPP;
|
return -ENOTSUPP;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!ring_size)
|
||||||
|
ring_size = HV_RING_SIZE * PAGE_SIZE;
|
||||||
|
|
||||||
|
/* Adjust ring size if necessary to have it page aligned */
|
||||||
|
ring_size = VMBUS_RING_SIZE(ring_size);
|
||||||
|
|
||||||
pdata = devm_kzalloc(&dev->device, sizeof(*pdata), GFP_KERNEL);
|
pdata = devm_kzalloc(&dev->device, sizeof(*pdata), GFP_KERNEL);
|
||||||
if (!pdata)
|
if (!pdata)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
ret = vmbus_alloc_ring(channel, HV_RING_SIZE * PAGE_SIZE,
|
ret = vmbus_alloc_ring(channel, ring_size, ring_size);
|
||||||
HV_RING_SIZE * PAGE_SIZE);
|
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
|
@@ -92,7 +92,6 @@ struct wdm_device {
|
|||||||
u16 wMaxCommand;
|
u16 wMaxCommand;
|
||||||
u16 wMaxPacketSize;
|
u16 wMaxPacketSize;
|
||||||
__le16 inum;
|
__le16 inum;
|
||||||
int reslength;
|
|
||||||
int length;
|
int length;
|
||||||
int read;
|
int read;
|
||||||
int count;
|
int count;
|
||||||
@@ -214,6 +213,11 @@ static void wdm_in_callback(struct urb *urb)
|
|||||||
if (desc->rerr == 0 && status != -EPIPE)
|
if (desc->rerr == 0 && status != -EPIPE)
|
||||||
desc->rerr = status;
|
desc->rerr = status;
|
||||||
|
|
||||||
|
if (length == 0) {
|
||||||
|
dev_dbg(&desc->intf->dev, "received ZLP\n");
|
||||||
|
goto skip_zlp;
|
||||||
|
}
|
||||||
|
|
||||||
if (length + desc->length > desc->wMaxCommand) {
|
if (length + desc->length > desc->wMaxCommand) {
|
||||||
/* The buffer would overflow */
|
/* The buffer would overflow */
|
||||||
set_bit(WDM_OVERFLOW, &desc->flags);
|
set_bit(WDM_OVERFLOW, &desc->flags);
|
||||||
@@ -222,18 +226,18 @@ static void wdm_in_callback(struct urb *urb)
|
|||||||
if (!test_bit(WDM_OVERFLOW, &desc->flags)) {
|
if (!test_bit(WDM_OVERFLOW, &desc->flags)) {
|
||||||
memmove(desc->ubuf + desc->length, desc->inbuf, length);
|
memmove(desc->ubuf + desc->length, desc->inbuf, length);
|
||||||
desc->length += length;
|
desc->length += length;
|
||||||
desc->reslength = length;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
skip_error:
|
skip_error:
|
||||||
|
|
||||||
if (desc->rerr) {
|
if (desc->rerr) {
|
||||||
/*
|
/*
|
||||||
* Since there was an error, userspace may decide to not read
|
* If there was a ZLP or an error, userspace may decide to not
|
||||||
* any data after poll'ing.
|
* read any data after poll'ing.
|
||||||
* We should respond to further attempts from the device to send
|
* We should respond to further attempts from the device to send
|
||||||
* data, so that we can get unstuck.
|
* data, so that we can get unstuck.
|
||||||
*/
|
*/
|
||||||
|
skip_zlp:
|
||||||
schedule_work(&desc->service_outs_intr);
|
schedule_work(&desc->service_outs_intr);
|
||||||
} else {
|
} else {
|
||||||
set_bit(WDM_READ, &desc->flags);
|
set_bit(WDM_READ, &desc->flags);
|
||||||
@@ -585,15 +589,6 @@ retry:
|
|||||||
goto retry;
|
goto retry;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!desc->reslength) { /* zero length read */
|
|
||||||
dev_dbg(&desc->intf->dev, "zero length - clearing WDM_READ\n");
|
|
||||||
clear_bit(WDM_READ, &desc->flags);
|
|
||||||
rv = service_outstanding_interrupt(desc);
|
|
||||||
spin_unlock_irq(&desc->iuspin);
|
|
||||||
if (rv < 0)
|
|
||||||
goto err;
|
|
||||||
goto retry;
|
|
||||||
}
|
|
||||||
cntr = desc->length;
|
cntr = desc->length;
|
||||||
spin_unlock_irq(&desc->iuspin);
|
spin_unlock_irq(&desc->iuspin);
|
||||||
}
|
}
|
||||||
@@ -1016,7 +1011,7 @@ static void service_interrupt_work(struct work_struct *work)
|
|||||||
|
|
||||||
spin_lock_irq(&desc->iuspin);
|
spin_lock_irq(&desc->iuspin);
|
||||||
service_outstanding_interrupt(desc);
|
service_outstanding_interrupt(desc);
|
||||||
if (!desc->resp_count) {
|
if (!desc->resp_count && (desc->length || desc->rerr)) {
|
||||||
set_bit(WDM_READ, &desc->flags);
|
set_bit(WDM_READ, &desc->flags);
|
||||||
wake_up(&desc->wait);
|
wake_up(&desc->wait);
|
||||||
}
|
}
|
||||||
|
@@ -20,6 +20,9 @@
|
|||||||
#include <linux/power_supply.h>
|
#include <linux/power_supply.h>
|
||||||
#include <linux/regulator/consumer.h>
|
#include <linux/regulator/consumer.h>
|
||||||
#include <linux/usb/role.h>
|
#include <linux/usb/role.h>
|
||||||
|
#include <linux/idr.h>
|
||||||
|
|
||||||
|
static DEFINE_IDA(usb_conn_ida);
|
||||||
|
|
||||||
#define USB_GPIO_DEB_MS 20 /* ms */
|
#define USB_GPIO_DEB_MS 20 /* ms */
|
||||||
#define USB_GPIO_DEB_US ((USB_GPIO_DEB_MS) * 1000) /* us */
|
#define USB_GPIO_DEB_US ((USB_GPIO_DEB_MS) * 1000) /* us */
|
||||||
@@ -29,6 +32,7 @@
|
|||||||
|
|
||||||
struct usb_conn_info {
|
struct usb_conn_info {
|
||||||
struct device *dev;
|
struct device *dev;
|
||||||
|
int conn_id; /* store the IDA-allocated ID */
|
||||||
struct usb_role_switch *role_sw;
|
struct usb_role_switch *role_sw;
|
||||||
enum usb_role last_role;
|
enum usb_role last_role;
|
||||||
struct regulator *vbus;
|
struct regulator *vbus;
|
||||||
@@ -160,7 +164,17 @@ static int usb_conn_psy_register(struct usb_conn_info *info)
|
|||||||
.of_node = dev->of_node,
|
.of_node = dev->of_node,
|
||||||
};
|
};
|
||||||
|
|
||||||
desc->name = "usb-charger";
|
info->conn_id = ida_alloc(&usb_conn_ida, GFP_KERNEL);
|
||||||
|
if (info->conn_id < 0)
|
||||||
|
return info->conn_id;
|
||||||
|
|
||||||
|
desc->name = devm_kasprintf(dev, GFP_KERNEL, "usb-charger-%d",
|
||||||
|
info->conn_id);
|
||||||
|
if (!desc->name) {
|
||||||
|
ida_free(&usb_conn_ida, info->conn_id);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
desc->properties = usb_charger_properties;
|
desc->properties = usb_charger_properties;
|
||||||
desc->num_properties = ARRAY_SIZE(usb_charger_properties);
|
desc->num_properties = ARRAY_SIZE(usb_charger_properties);
|
||||||
desc->get_property = usb_charger_get_property;
|
desc->get_property = usb_charger_get_property;
|
||||||
@@ -168,8 +182,10 @@ static int usb_conn_psy_register(struct usb_conn_info *info)
|
|||||||
cfg.drv_data = info;
|
cfg.drv_data = info;
|
||||||
|
|
||||||
info->charger = devm_power_supply_register(dev, desc, &cfg);
|
info->charger = devm_power_supply_register(dev, desc, &cfg);
|
||||||
if (IS_ERR(info->charger))
|
if (IS_ERR(info->charger)) {
|
||||||
dev_err(dev, "Unable to register charger\n");
|
dev_err(dev, "Unable to register charger %d\n", info->conn_id);
|
||||||
|
ida_free(&usb_conn_ida, info->conn_id);
|
||||||
|
}
|
||||||
|
|
||||||
return PTR_ERR_OR_ZERO(info->charger);
|
return PTR_ERR_OR_ZERO(info->charger);
|
||||||
}
|
}
|
||||||
@@ -277,6 +293,9 @@ static void usb_conn_remove(struct platform_device *pdev)
|
|||||||
|
|
||||||
cancel_delayed_work_sync(&info->dw_det);
|
cancel_delayed_work_sync(&info->dw_det);
|
||||||
|
|
||||||
|
if (info->charger)
|
||||||
|
ida_free(&usb_conn_ida, info->conn_id);
|
||||||
|
|
||||||
if (info->last_role == USB_ROLE_HOST && info->vbus)
|
if (info->last_role == USB_ROLE_HOST && info->vbus)
|
||||||
regulator_disable(info->vbus);
|
regulator_disable(info->vbus);
|
||||||
|
|
||||||
|
@@ -695,15 +695,16 @@ struct usb_device *usb_alloc_dev(struct usb_device *parent,
|
|||||||
device_set_of_node_from_dev(&dev->dev, bus->sysdev);
|
device_set_of_node_from_dev(&dev->dev, bus->sysdev);
|
||||||
dev_set_name(&dev->dev, "usb%d", bus->busnum);
|
dev_set_name(&dev->dev, "usb%d", bus->busnum);
|
||||||
} else {
|
} else {
|
||||||
|
int n;
|
||||||
|
|
||||||
/* match any labeling on the hubs; it's one-based */
|
/* match any labeling on the hubs; it's one-based */
|
||||||
if (parent->devpath[0] == '0') {
|
if (parent->devpath[0] == '0') {
|
||||||
snprintf(dev->devpath, sizeof dev->devpath,
|
n = snprintf(dev->devpath, sizeof(dev->devpath), "%d", port1);
|
||||||
"%d", port1);
|
|
||||||
/* Root ports are not counted in route string */
|
/* Root ports are not counted in route string */
|
||||||
dev->route = 0;
|
dev->route = 0;
|
||||||
} else {
|
} else {
|
||||||
snprintf(dev->devpath, sizeof dev->devpath,
|
n = snprintf(dev->devpath, sizeof(dev->devpath), "%s.%d",
|
||||||
"%s.%d", parent->devpath, port1);
|
parent->devpath, port1);
|
||||||
/* Route string assumes hubs have less than 16 ports */
|
/* Route string assumes hubs have less than 16 ports */
|
||||||
if (port1 < 15)
|
if (port1 < 15)
|
||||||
dev->route = parent->route +
|
dev->route = parent->route +
|
||||||
@@ -712,6 +713,11 @@ struct usb_device *usb_alloc_dev(struct usb_device *parent,
|
|||||||
dev->route = parent->route +
|
dev->route = parent->route +
|
||||||
(15 << ((parent->level - 1)*4));
|
(15 << ((parent->level - 1)*4));
|
||||||
}
|
}
|
||||||
|
if (n >= sizeof(dev->devpath)) {
|
||||||
|
usb_put_hcd(bus_to_hcd(bus));
|
||||||
|
usb_put_dev(dev);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
dev->dev.parent = &parent->dev;
|
dev->dev.parent = &parent->dev;
|
||||||
dev_set_name(&dev->dev, "%d-%s", bus->busnum, dev->devpath);
|
dev_set_name(&dev->dev, "%d-%s", bus->busnum, dev->devpath);
|
||||||
|
@@ -4601,6 +4601,12 @@ static int dwc2_hsotg_udc_stop(struct usb_gadget *gadget)
|
|||||||
if (!hsotg)
|
if (!hsotg)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
|
|
||||||
|
/* Exit clock gating when driver is stopped. */
|
||||||
|
if (hsotg->params.power_down == DWC2_POWER_DOWN_PARAM_NONE &&
|
||||||
|
hsotg->bus_suspended && !hsotg->params.no_clock_gating) {
|
||||||
|
dwc2_gadget_exit_clock_gating(hsotg, 0);
|
||||||
|
}
|
||||||
|
|
||||||
/* all endpoints should be shutdown */
|
/* all endpoints should be shutdown */
|
||||||
for (ep = 1; ep < hsotg->num_of_eps; ep++) {
|
for (ep = 1; ep < hsotg->num_of_eps; ep++) {
|
||||||
if (hsotg->eps_in[ep])
|
if (hsotg->eps_in[ep])
|
||||||
|
@@ -1297,14 +1297,14 @@ static struct se_portal_group *usbg_make_tpg(struct se_wwn *wwn,
|
|||||||
struct usbg_tport *tport = container_of(wwn, struct usbg_tport,
|
struct usbg_tport *tport = container_of(wwn, struct usbg_tport,
|
||||||
tport_wwn);
|
tport_wwn);
|
||||||
struct usbg_tpg *tpg;
|
struct usbg_tpg *tpg;
|
||||||
unsigned long tpgt;
|
u16 tpgt;
|
||||||
int ret;
|
int ret;
|
||||||
struct f_tcm_opts *opts;
|
struct f_tcm_opts *opts;
|
||||||
unsigned i;
|
unsigned i;
|
||||||
|
|
||||||
if (strstr(name, "tpgt_") != name)
|
if (strstr(name, "tpgt_") != name)
|
||||||
return ERR_PTR(-EINVAL);
|
return ERR_PTR(-EINVAL);
|
||||||
if (kstrtoul(name + 5, 0, &tpgt) || tpgt > UINT_MAX)
|
if (kstrtou16(name + 5, 0, &tpgt))
|
||||||
return ERR_PTR(-EINVAL);
|
return ERR_PTR(-EINVAL);
|
||||||
ret = -ENODEV;
|
ret = -ENODEV;
|
||||||
mutex_lock(&tpg_instances_lock);
|
mutex_lock(&tpg_instances_lock);
|
||||||
|
@@ -323,6 +323,10 @@ static int dp_altmode_vdm(struct typec_altmode *alt,
|
|||||||
break;
|
break;
|
||||||
case CMDT_RSP_NAK:
|
case CMDT_RSP_NAK:
|
||||||
switch (cmd) {
|
switch (cmd) {
|
||||||
|
case DP_CMD_STATUS_UPDATE:
|
||||||
|
if (typec_altmode_exit(alt))
|
||||||
|
dev_err(&dp->alt->dev, "Exit Mode Failed!\n");
|
||||||
|
break;
|
||||||
case DP_CMD_CONFIGURE:
|
case DP_CMD_CONFIGURE:
|
||||||
dp->data.conf = 0;
|
dp->data.conf = 0;
|
||||||
ret = dp_altmode_configured(dp);
|
ret = dp_altmode_configured(dp);
|
||||||
|
@@ -214,7 +214,7 @@ int typec_switch_set(struct typec_switch *sw,
|
|||||||
sw_dev = sw->sw_devs[i];
|
sw_dev = sw->sw_devs[i];
|
||||||
|
|
||||||
ret = sw_dev->set(sw_dev, orientation);
|
ret = sw_dev->set(sw_dev, orientation);
|
||||||
if (ret)
|
if (ret && ret != -EOPNOTSUPP)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -378,7 +378,7 @@ int typec_mux_set(struct typec_mux *mux, struct typec_mux_state *state)
|
|||||||
mux_dev = mux->mux_devs[i];
|
mux_dev = mux->mux_devs[i];
|
||||||
|
|
||||||
ret = mux_dev->set(mux_dev, state);
|
ret = mux_dev->set(mux_dev, state);
|
||||||
if (ret)
|
if (ret && ret != -EOPNOTSUPP)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -82,6 +82,15 @@ static int dummycon_blank(struct vc_data *vc, int blank, int mode_switch)
|
|||||||
/* Redraw, so that we get putc(s) for output done while blanked */
|
/* Redraw, so that we get putc(s) for output done while blanked */
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static bool dummycon_switch(struct vc_data *vc)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* Redraw, so that we get putc(s) for output done while switched
|
||||||
|
* away. Informs deferred consoles to take over the display.
|
||||||
|
*/
|
||||||
|
return true;
|
||||||
|
}
|
||||||
#else
|
#else
|
||||||
static void dummycon_putc(struct vc_data *vc, int c, int ypos, int xpos) { }
|
static void dummycon_putc(struct vc_data *vc, int c, int ypos, int xpos) { }
|
||||||
static void dummycon_putcs(struct vc_data *vc, const unsigned short *s,
|
static void dummycon_putcs(struct vc_data *vc, const unsigned short *s,
|
||||||
@@ -90,6 +99,10 @@ static int dummycon_blank(struct vc_data *vc, int blank, int mode_switch)
|
|||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
static bool dummycon_switch(struct vc_data *vc)
|
||||||
|
{
|
||||||
|
return false;
|
||||||
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static const char *dummycon_startup(void)
|
static const char *dummycon_startup(void)
|
||||||
@@ -97,7 +110,7 @@ static const char *dummycon_startup(void)
|
|||||||
return "dummy device";
|
return "dummy device";
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dummycon_init(struct vc_data *vc, int init)
|
static void dummycon_init(struct vc_data *vc, bool init)
|
||||||
{
|
{
|
||||||
vc->vc_can_do_color = 1;
|
vc->vc_can_do_color = 1;
|
||||||
if (init) {
|
if (init) {
|
||||||
@@ -108,8 +121,8 @@ static void dummycon_init(struct vc_data *vc, int init)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void dummycon_deinit(struct vc_data *vc) { }
|
static void dummycon_deinit(struct vc_data *vc) { }
|
||||||
static void dummycon_clear(struct vc_data *vc, int sy, int sx, int height,
|
static void dummycon_clear(struct vc_data *vc, unsigned int sy, unsigned int sx,
|
||||||
int width) { }
|
unsigned int width) { }
|
||||||
static void dummycon_cursor(struct vc_data *vc, int mode) { }
|
static void dummycon_cursor(struct vc_data *vc, int mode) { }
|
||||||
|
|
||||||
static bool dummycon_scroll(struct vc_data *vc, unsigned int top,
|
static bool dummycon_scroll(struct vc_data *vc, unsigned int top,
|
||||||
@@ -119,11 +132,6 @@ static bool dummycon_scroll(struct vc_data *vc, unsigned int top,
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dummycon_switch(struct vc_data *vc)
|
|
||||||
{
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The console `switch' structure for the dummy console
|
* The console `switch' structure for the dummy console
|
||||||
*
|
*
|
||||||
|
@@ -352,7 +352,7 @@ static const char *mdacon_startup(void)
|
|||||||
return "MDA-2";
|
return "MDA-2";
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mdacon_init(struct vc_data *c, int init)
|
static void mdacon_init(struct vc_data *c, bool init)
|
||||||
{
|
{
|
||||||
c->vc_complement_mask = 0x0800; /* reverse video */
|
c->vc_complement_mask = 0x0800; /* reverse video */
|
||||||
c->vc_display_fg = &mda_display_fg;
|
c->vc_display_fg = &mda_display_fg;
|
||||||
@@ -442,26 +442,21 @@ static void mdacon_putcs(struct vc_data *c, const unsigned short *s,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mdacon_clear(struct vc_data *c, int y, int x,
|
static void mdacon_clear(struct vc_data *c, unsigned int y, unsigned int x,
|
||||||
int height, int width)
|
unsigned int width)
|
||||||
{
|
{
|
||||||
u16 *dest = mda_addr(x, y);
|
u16 *dest = mda_addr(x, y);
|
||||||
u16 eattr = mda_convert_attr(c->vc_video_erase_char);
|
u16 eattr = mda_convert_attr(c->vc_video_erase_char);
|
||||||
|
|
||||||
if (width <= 0 || height <= 0)
|
if (width <= 0)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (x==0 && width==mda_num_columns) {
|
scr_memsetw(dest, eattr, width * 2);
|
||||||
scr_memsetw(dest, eattr, height*width*2);
|
|
||||||
} else {
|
|
||||||
for (; height > 0; height--, dest+=mda_num_columns)
|
|
||||||
scr_memsetw(dest, eattr, width*2);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int mdacon_switch(struct vc_data *c)
|
static bool mdacon_switch(struct vc_data *c)
|
||||||
{
|
{
|
||||||
return 1; /* redrawing needed */
|
return true; /* redrawing needed */
|
||||||
}
|
}
|
||||||
|
|
||||||
static int mdacon_blank(struct vc_data *c, int blank, int mode_switch)
|
static int mdacon_blank(struct vc_data *c, int blank, int mode_switch)
|
||||||
|
@@ -324,7 +324,7 @@ out_unmap:
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void newport_init(struct vc_data *vc, int init)
|
static void newport_init(struct vc_data *vc, bool init)
|
||||||
{
|
{
|
||||||
int cols, rows;
|
int cols, rows;
|
||||||
|
|
||||||
@@ -346,12 +346,12 @@ static void newport_deinit(struct vc_data *c)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void newport_clear(struct vc_data *vc, int sy, int sx, int height,
|
static void newport_clear(struct vc_data *vc, unsigned int sy, unsigned int sx,
|
||||||
int width)
|
unsigned int width)
|
||||||
{
|
{
|
||||||
int xend = ((sx + width) << 3) - 1;
|
int xend = ((sx + width) << 3) - 1;
|
||||||
int ystart = ((sy << 4) + topscan) & 0x3ff;
|
int ystart = ((sy << 4) + topscan) & 0x3ff;
|
||||||
int yend = (((sy + height) << 4) + topscan - 1) & 0x3ff;
|
int yend = (((sy + 1) << 4) + topscan - 1) & 0x3ff;
|
||||||
|
|
||||||
if (logo_active)
|
if (logo_active)
|
||||||
return;
|
return;
|
||||||
@@ -462,7 +462,7 @@ static void newport_cursor(struct vc_data *vc, int mode)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int newport_switch(struct vc_data *vc)
|
static bool newport_switch(struct vc_data *vc)
|
||||||
{
|
{
|
||||||
static int logo_drawn = 0;
|
static int logo_drawn = 0;
|
||||||
|
|
||||||
@@ -476,7 +476,7 @@ static int newport_switch(struct vc_data *vc)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return 1;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int newport_blank(struct vc_data *c, int blank, int mode_switch)
|
static int newport_blank(struct vc_data *c, int blank, int mode_switch)
|
||||||
|
@@ -273,7 +273,7 @@ static int sticon_font_set(struct vc_data *vc, struct console_font *font,
|
|||||||
return sticon_set_font(vc, font, vpitch);
|
return sticon_set_font(vc, font, vpitch);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void sticon_init(struct vc_data *c, int init)
|
static void sticon_init(struct vc_data *c, bool init)
|
||||||
{
|
{
|
||||||
struct sti_struct *sti = sticon_sti;
|
struct sti_struct *sti = sticon_sti;
|
||||||
int vc_cols, vc_rows;
|
int vc_cols, vc_rows;
|
||||||
@@ -300,19 +300,19 @@ static void sticon_deinit(struct vc_data *c)
|
|||||||
sticon_set_def_font(i);
|
sticon_set_def_font(i);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void sticon_clear(struct vc_data *conp, int sy, int sx, int height,
|
static void sticon_clear(struct vc_data *conp, unsigned int sy, unsigned int sx,
|
||||||
int width)
|
unsigned int width)
|
||||||
{
|
{
|
||||||
if (!height || !width)
|
if (!width)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
sti_clear(sticon_sti, sy, sx, height, width,
|
sti_clear(sticon_sti, sy, sx, 1, width,
|
||||||
conp->vc_video_erase_char, font_data[conp->vc_num]);
|
conp->vc_video_erase_char, font_data[conp->vc_num]);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int sticon_switch(struct vc_data *conp)
|
static bool sticon_switch(struct vc_data *conp)
|
||||||
{
|
{
|
||||||
return 1; /* needs refreshing */
|
return true; /* needs refreshing */
|
||||||
}
|
}
|
||||||
|
|
||||||
static int sticon_blank(struct vc_data *c, int blank, int mode_switch)
|
static int sticon_blank(struct vc_data *c, int blank, int mode_switch)
|
||||||
|
@@ -332,7 +332,7 @@ static const char *vgacon_startup(void)
|
|||||||
return display_desc;
|
return display_desc;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void vgacon_init(struct vc_data *c, int init)
|
static void vgacon_init(struct vc_data *c, bool init)
|
||||||
{
|
{
|
||||||
struct uni_pagedict *p;
|
struct uni_pagedict *p;
|
||||||
|
|
||||||
@@ -349,7 +349,7 @@ static void vgacon_init(struct vc_data *c, int init)
|
|||||||
c->vc_scan_lines = vga_scan_lines;
|
c->vc_scan_lines = vga_scan_lines;
|
||||||
c->vc_font.height = c->vc_cell_height = vga_video_font_height;
|
c->vc_font.height = c->vc_cell_height = vga_video_font_height;
|
||||||
|
|
||||||
/* set dimensions manually if init != 0 since vc_resize() will fail */
|
/* set dimensions manually if init is true since vc_resize() will fail */
|
||||||
if (init) {
|
if (init) {
|
||||||
c->vc_cols = vga_video_num_columns;
|
c->vc_cols = vga_video_num_columns;
|
||||||
c->vc_rows = vga_video_num_lines;
|
c->vc_rows = vga_video_num_lines;
|
||||||
@@ -585,7 +585,7 @@ static void vgacon_doresize(struct vc_data *c,
|
|||||||
raw_spin_unlock_irqrestore(&vga_lock, flags);
|
raw_spin_unlock_irqrestore(&vga_lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int vgacon_switch(struct vc_data *c)
|
static bool vgacon_switch(struct vc_data *c)
|
||||||
{
|
{
|
||||||
int x = c->vc_cols * VGA_FONTWIDTH;
|
int x = c->vc_cols * VGA_FONTWIDTH;
|
||||||
int y = c->vc_rows * c->vc_cell_height;
|
int y = c->vc_rows * c->vc_cell_height;
|
||||||
@@ -614,7 +614,7 @@ static int vgacon_switch(struct vc_data *c)
|
|||||||
vgacon_doresize(c, c->vc_cols, c->vc_rows);
|
vgacon_doresize(c, c->vc_cols, c->vc_rows);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0; /* Redrawing not needed */
|
return false; /* Redrawing not needed */
|
||||||
}
|
}
|
||||||
|
|
||||||
static void vga_set_palette(struct vc_data *vc, const unsigned char *table)
|
static void vga_set_palette(struct vc_data *vc, const unsigned char *table)
|
||||||
@@ -1156,8 +1156,8 @@ static bool vgacon_scroll(struct vc_data *c, unsigned int t, unsigned int b,
|
|||||||
* The console `switch' structure for the VGA based console
|
* The console `switch' structure for the VGA based console
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void vgacon_clear(struct vc_data *vc, int sy, int sx, int height,
|
static void vgacon_clear(struct vc_data *vc, unsigned int sy, unsigned int sx,
|
||||||
int width) { }
|
unsigned int width) { }
|
||||||
static void vgacon_putc(struct vc_data *vc, int c, int ypos, int xpos) { }
|
static void vgacon_putc(struct vc_data *vc, int c, int ypos, int xpos) { }
|
||||||
static void vgacon_putcs(struct vc_data *vc, const unsigned short *s,
|
static void vgacon_putcs(struct vc_data *vc, const unsigned short *s,
|
||||||
int count, int ypos, int xpos) { }
|
int count, int ypos, int xpos) { }
|
||||||
|
@@ -993,7 +993,7 @@ static const char *fbcon_startup(void)
|
|||||||
return display_desc;
|
return display_desc;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void fbcon_init(struct vc_data *vc, int init)
|
static void fbcon_init(struct vc_data *vc, bool init)
|
||||||
{
|
{
|
||||||
struct fb_info *info;
|
struct fb_info *info;
|
||||||
struct fbcon_ops *ops;
|
struct fbcon_ops *ops;
|
||||||
@@ -1240,8 +1240,8 @@ finished:
|
|||||||
* restriction is simplicity & efficiency at the moment.
|
* restriction is simplicity & efficiency at the moment.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static void fbcon_clear(struct vc_data *vc, int sy, int sx, int height,
|
static void __fbcon_clear(struct vc_data *vc, unsigned int sy, unsigned int sx,
|
||||||
int width)
|
unsigned int height, unsigned int width)
|
||||||
{
|
{
|
||||||
struct fb_info *info = fbcon_info_from_console(vc->vc_num);
|
struct fb_info *info = fbcon_info_from_console(vc->vc_num);
|
||||||
struct fbcon_ops *ops = info->fbcon_par;
|
struct fbcon_ops *ops = info->fbcon_par;
|
||||||
@@ -1280,6 +1280,12 @@ static void fbcon_clear(struct vc_data *vc, int sy, int sx, int height,
|
|||||||
ops->clear(vc, info, real_y(p, sy), sx, height, width, fg, bg);
|
ops->clear(vc, info, real_y(p, sy), sx, height, width, fg, bg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void fbcon_clear(struct vc_data *vc, unsigned int sy, unsigned int sx,
|
||||||
|
unsigned int width)
|
||||||
|
{
|
||||||
|
__fbcon_clear(vc, sy, sx, 1, width);
|
||||||
|
}
|
||||||
|
|
||||||
static void fbcon_putcs(struct vc_data *vc, const unsigned short *s,
|
static void fbcon_putcs(struct vc_data *vc, const unsigned short *s,
|
||||||
int count, int ypos, int xpos)
|
int count, int ypos, int xpos)
|
||||||
{
|
{
|
||||||
@@ -1767,7 +1773,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
case SCROLL_MOVE:
|
case SCROLL_MOVE:
|
||||||
fbcon_redraw_blit(vc, info, p, t, b - t - count,
|
fbcon_redraw_blit(vc, info, p, t, b - t - count,
|
||||||
count);
|
count);
|
||||||
fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
||||||
scr_memsetw((unsigned short *) (vc->vc_origin +
|
scr_memsetw((unsigned short *) (vc->vc_origin +
|
||||||
vc->vc_size_row *
|
vc->vc_size_row *
|
||||||
(b - count)),
|
(b - count)),
|
||||||
@@ -1790,7 +1796,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
b - t - count, vc->vc_cols);
|
b - t - count, vc->vc_cols);
|
||||||
else
|
else
|
||||||
goto redraw_up;
|
goto redraw_up;
|
||||||
fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_PAN_REDRAW:
|
case SCROLL_PAN_REDRAW:
|
||||||
@@ -1808,7 +1814,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
vc->vc_rows - b, b);
|
vc->vc_rows - b, b);
|
||||||
} else
|
} else
|
||||||
fbcon_redraw_move(vc, p, t + count, b - t - count, t);
|
fbcon_redraw_move(vc, p, t + count, b - t - count, t);
|
||||||
fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_PAN_MOVE:
|
case SCROLL_PAN_MOVE:
|
||||||
@@ -1831,14 +1837,14 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
b - t - count, vc->vc_cols);
|
b - t - count, vc->vc_cols);
|
||||||
else
|
else
|
||||||
goto redraw_up;
|
goto redraw_up;
|
||||||
fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_REDRAW:
|
case SCROLL_REDRAW:
|
||||||
redraw_up:
|
redraw_up:
|
||||||
fbcon_redraw(vc, t, b - t - count,
|
fbcon_redraw(vc, t, b - t - count,
|
||||||
count * vc->vc_cols);
|
count * vc->vc_cols);
|
||||||
fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, b - count, 0, count, vc->vc_cols);
|
||||||
scr_memsetw((unsigned short *) (vc->vc_origin +
|
scr_memsetw((unsigned short *) (vc->vc_origin +
|
||||||
vc->vc_size_row *
|
vc->vc_size_row *
|
||||||
(b - count)),
|
(b - count)),
|
||||||
@@ -1855,7 +1861,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
case SCROLL_MOVE:
|
case SCROLL_MOVE:
|
||||||
fbcon_redraw_blit(vc, info, p, b - 1, b - t - count,
|
fbcon_redraw_blit(vc, info, p, b - 1, b - t - count,
|
||||||
-count);
|
-count);
|
||||||
fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
||||||
scr_memsetw((unsigned short *) (vc->vc_origin +
|
scr_memsetw((unsigned short *) (vc->vc_origin +
|
||||||
vc->vc_size_row *
|
vc->vc_size_row *
|
||||||
t),
|
t),
|
||||||
@@ -1878,7 +1884,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
b - t - count, vc->vc_cols);
|
b - t - count, vc->vc_cols);
|
||||||
else
|
else
|
||||||
goto redraw_down;
|
goto redraw_down;
|
||||||
fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_PAN_MOVE:
|
case SCROLL_PAN_MOVE:
|
||||||
@@ -1900,7 +1906,7 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
b - t - count, vc->vc_cols);
|
b - t - count, vc->vc_cols);
|
||||||
else
|
else
|
||||||
goto redraw_down;
|
goto redraw_down;
|
||||||
fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_PAN_REDRAW:
|
case SCROLL_PAN_REDRAW:
|
||||||
@@ -1917,14 +1923,14 @@ static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b,
|
|||||||
fbcon_redraw_move(vc, p, count, t, 0);
|
fbcon_redraw_move(vc, p, count, t, 0);
|
||||||
} else
|
} else
|
||||||
fbcon_redraw_move(vc, p, t, b - t - count, t + count);
|
fbcon_redraw_move(vc, p, t, b - t - count, t + count);
|
||||||
fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case SCROLL_REDRAW:
|
case SCROLL_REDRAW:
|
||||||
redraw_down:
|
redraw_down:
|
||||||
fbcon_redraw(vc, b - 1, b - t - count,
|
fbcon_redraw(vc, b - 1, b - t - count,
|
||||||
-count * vc->vc_cols);
|
-count * vc->vc_cols);
|
||||||
fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
__fbcon_clear(vc, t, 0, count, vc->vc_cols);
|
||||||
scr_memsetw((unsigned short *) (vc->vc_origin +
|
scr_memsetw((unsigned short *) (vc->vc_origin +
|
||||||
vc->vc_size_row *
|
vc->vc_size_row *
|
||||||
t),
|
t),
|
||||||
@@ -2066,7 +2072,7 @@ static int fbcon_resize(struct vc_data *vc, unsigned int width,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int fbcon_switch(struct vc_data *vc)
|
static bool fbcon_switch(struct vc_data *vc)
|
||||||
{
|
{
|
||||||
struct fb_info *info, *old_info = NULL;
|
struct fb_info *info, *old_info = NULL;
|
||||||
struct fbcon_ops *ops;
|
struct fbcon_ops *ops;
|
||||||
@@ -2188,9 +2194,9 @@ static int fbcon_switch(struct vc_data *vc)
|
|||||||
vc->vc_origin + vc->vc_size_row * vc->vc_top,
|
vc->vc_origin + vc->vc_size_row * vc->vc_top,
|
||||||
vc->vc_size_row * (vc->vc_bottom -
|
vc->vc_size_row * (vc->vc_bottom -
|
||||||
vc->vc_top) / 2);
|
vc->vc_top) / 2);
|
||||||
return 0;
|
return false;
|
||||||
}
|
}
|
||||||
return 1;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void fbcon_generic_blank(struct vc_data *vc, struct fb_info *info,
|
static void fbcon_generic_blank(struct vc_data *vc, struct fb_info *info,
|
||||||
@@ -2203,7 +2209,7 @@ static void fbcon_generic_blank(struct vc_data *vc, struct fb_info *info,
|
|||||||
|
|
||||||
oldc = vc->vc_video_erase_char;
|
oldc = vc->vc_video_erase_char;
|
||||||
vc->vc_video_erase_char &= charmask;
|
vc->vc_video_erase_char &= charmask;
|
||||||
fbcon_clear(vc, 0, 0, vc->vc_rows, vc->vc_cols);
|
__fbcon_clear(vc, 0, 0, vc->vc_rows, vc->vc_cols);
|
||||||
vc->vc_video_erase_char = oldc;
|
vc->vc_video_erase_char = oldc;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -2148,7 +2148,6 @@ static int load_global_roots_objectid(struct btrfs_root *tree_root,
|
|||||||
found = true;
|
found = true;
|
||||||
root = read_tree_root_path(tree_root, path, &key);
|
root = read_tree_root_path(tree_root, path, &key);
|
||||||
if (IS_ERR(root)) {
|
if (IS_ERR(root)) {
|
||||||
if (!btrfs_test_opt(fs_info, IGNOREBADROOTS))
|
|
||||||
ret = PTR_ERR(root);
|
ret = PTR_ERR(root);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@@ -8754,6 +8754,7 @@ static int btrfs_rename_exchange(struct inode *old_dir,
|
|||||||
int ret;
|
int ret;
|
||||||
int ret2;
|
int ret2;
|
||||||
bool need_abort = false;
|
bool need_abort = false;
|
||||||
|
bool logs_pinned = false;
|
||||||
struct fscrypt_name old_fname, new_fname;
|
struct fscrypt_name old_fname, new_fname;
|
||||||
struct fscrypt_str *old_name, *new_name;
|
struct fscrypt_str *old_name, *new_name;
|
||||||
|
|
||||||
@@ -8877,6 +8878,31 @@ static int btrfs_rename_exchange(struct inode *old_dir,
|
|||||||
inode_inc_iversion(new_inode);
|
inode_inc_iversion(new_inode);
|
||||||
simple_rename_timestamp(old_dir, old_dentry, new_dir, new_dentry);
|
simple_rename_timestamp(old_dir, old_dentry, new_dir, new_dentry);
|
||||||
|
|
||||||
|
if (old_ino != BTRFS_FIRST_FREE_OBJECTID &&
|
||||||
|
new_ino != BTRFS_FIRST_FREE_OBJECTID) {
|
||||||
|
/*
|
||||||
|
* If we are renaming in the same directory (and it's not for
|
||||||
|
* root entries) pin the log early to prevent any concurrent
|
||||||
|
* task from logging the directory after we removed the old
|
||||||
|
* entries and before we add the new entries, otherwise that
|
||||||
|
* task can sync a log without any entry for the inodes we are
|
||||||
|
* renaming and therefore replaying that log, if a power failure
|
||||||
|
* happens after syncing the log, would result in deleting the
|
||||||
|
* inodes.
|
||||||
|
*
|
||||||
|
* If the rename affects two different directories, we want to
|
||||||
|
* make sure the that there's no log commit that contains
|
||||||
|
* updates for only one of the directories but not for the
|
||||||
|
* other.
|
||||||
|
*
|
||||||
|
* If we are renaming an entry for a root, we don't care about
|
||||||
|
* log updates since we called btrfs_set_log_full_commit().
|
||||||
|
*/
|
||||||
|
btrfs_pin_log_trans(root);
|
||||||
|
btrfs_pin_log_trans(dest);
|
||||||
|
logs_pinned = true;
|
||||||
|
}
|
||||||
|
|
||||||
if (old_dentry->d_parent != new_dentry->d_parent) {
|
if (old_dentry->d_parent != new_dentry->d_parent) {
|
||||||
btrfs_record_unlink_dir(trans, BTRFS_I(old_dir),
|
btrfs_record_unlink_dir(trans, BTRFS_I(old_dir),
|
||||||
BTRFS_I(old_inode), true);
|
BTRFS_I(old_inode), true);
|
||||||
@@ -8934,30 +8960,23 @@ static int btrfs_rename_exchange(struct inode *old_dir,
|
|||||||
BTRFS_I(new_inode)->dir_index = new_idx;
|
BTRFS_I(new_inode)->dir_index = new_idx;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Now pin the logs of the roots. We do it to ensure that no other task
|
* Do the log updates for all inodes.
|
||||||
* can sync the logs while we are in progress with the rename, because
|
*
|
||||||
* that could result in an inconsistency in case any of the inodes that
|
* If either entry is for a root we don't need to update the logs since
|
||||||
* are part of this rename operation were logged before.
|
* we've called btrfs_set_log_full_commit() before.
|
||||||
*/
|
*/
|
||||||
if (old_ino != BTRFS_FIRST_FREE_OBJECTID)
|
if (logs_pinned) {
|
||||||
btrfs_pin_log_trans(root);
|
|
||||||
if (new_ino != BTRFS_FIRST_FREE_OBJECTID)
|
|
||||||
btrfs_pin_log_trans(dest);
|
|
||||||
|
|
||||||
/* Do the log updates for all inodes. */
|
|
||||||
if (old_ino != BTRFS_FIRST_FREE_OBJECTID)
|
|
||||||
btrfs_log_new_name(trans, old_dentry, BTRFS_I(old_dir),
|
btrfs_log_new_name(trans, old_dentry, BTRFS_I(old_dir),
|
||||||
old_rename_ctx.index, new_dentry->d_parent);
|
old_rename_ctx.index, new_dentry->d_parent);
|
||||||
if (new_ino != BTRFS_FIRST_FREE_OBJECTID)
|
|
||||||
btrfs_log_new_name(trans, new_dentry, BTRFS_I(new_dir),
|
btrfs_log_new_name(trans, new_dentry, BTRFS_I(new_dir),
|
||||||
new_rename_ctx.index, old_dentry->d_parent);
|
new_rename_ctx.index, old_dentry->d_parent);
|
||||||
|
}
|
||||||
|
|
||||||
/* Now unpin the logs. */
|
|
||||||
if (old_ino != BTRFS_FIRST_FREE_OBJECTID)
|
|
||||||
btrfs_end_log_trans(root);
|
|
||||||
if (new_ino != BTRFS_FIRST_FREE_OBJECTID)
|
|
||||||
btrfs_end_log_trans(dest);
|
|
||||||
out_fail:
|
out_fail:
|
||||||
|
if (logs_pinned) {
|
||||||
|
btrfs_end_log_trans(root);
|
||||||
|
btrfs_end_log_trans(dest);
|
||||||
|
}
|
||||||
ret2 = btrfs_end_transaction(trans);
|
ret2 = btrfs_end_transaction(trans);
|
||||||
ret = ret ? ret : ret2;
|
ret = ret ? ret : ret2;
|
||||||
out_notrans:
|
out_notrans:
|
||||||
@@ -9007,6 +9026,7 @@ static int btrfs_rename(struct mnt_idmap *idmap,
|
|||||||
int ret2;
|
int ret2;
|
||||||
u64 old_ino = btrfs_ino(BTRFS_I(old_inode));
|
u64 old_ino = btrfs_ino(BTRFS_I(old_inode));
|
||||||
struct fscrypt_name old_fname, new_fname;
|
struct fscrypt_name old_fname, new_fname;
|
||||||
|
bool logs_pinned = false;
|
||||||
|
|
||||||
if (btrfs_ino(BTRFS_I(new_dir)) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)
|
if (btrfs_ino(BTRFS_I(new_dir)) == BTRFS_EMPTY_SUBVOL_DIR_OBJECTID)
|
||||||
return -EPERM;
|
return -EPERM;
|
||||||
@@ -9141,6 +9161,29 @@ static int btrfs_rename(struct mnt_idmap *idmap,
|
|||||||
inode_inc_iversion(old_inode);
|
inode_inc_iversion(old_inode);
|
||||||
simple_rename_timestamp(old_dir, old_dentry, new_dir, new_dentry);
|
simple_rename_timestamp(old_dir, old_dentry, new_dir, new_dentry);
|
||||||
|
|
||||||
|
if (old_ino != BTRFS_FIRST_FREE_OBJECTID) {
|
||||||
|
/*
|
||||||
|
* If we are renaming in the same directory (and it's not a
|
||||||
|
* root entry) pin the log to prevent any concurrent task from
|
||||||
|
* logging the directory after we removed the old entry and
|
||||||
|
* before we add the new entry, otherwise that task can sync
|
||||||
|
* a log without any entry for the inode we are renaming and
|
||||||
|
* therefore replaying that log, if a power failure happens
|
||||||
|
* after syncing the log, would result in deleting the inode.
|
||||||
|
*
|
||||||
|
* If the rename affects two different directories, we want to
|
||||||
|
* make sure the that there's no log commit that contains
|
||||||
|
* updates for only one of the directories but not for the
|
||||||
|
* other.
|
||||||
|
*
|
||||||
|
* If we are renaming an entry for a root, we don't care about
|
||||||
|
* log updates since we called btrfs_set_log_full_commit().
|
||||||
|
*/
|
||||||
|
btrfs_pin_log_trans(root);
|
||||||
|
btrfs_pin_log_trans(dest);
|
||||||
|
logs_pinned = true;
|
||||||
|
}
|
||||||
|
|
||||||
if (old_dentry->d_parent != new_dentry->d_parent)
|
if (old_dentry->d_parent != new_dentry->d_parent)
|
||||||
btrfs_record_unlink_dir(trans, BTRFS_I(old_dir),
|
btrfs_record_unlink_dir(trans, BTRFS_I(old_dir),
|
||||||
BTRFS_I(old_inode), true);
|
BTRFS_I(old_inode), true);
|
||||||
@@ -9189,7 +9232,7 @@ static int btrfs_rename(struct mnt_idmap *idmap,
|
|||||||
if (old_inode->i_nlink == 1)
|
if (old_inode->i_nlink == 1)
|
||||||
BTRFS_I(old_inode)->dir_index = index;
|
BTRFS_I(old_inode)->dir_index = index;
|
||||||
|
|
||||||
if (old_ino != BTRFS_FIRST_FREE_OBJECTID)
|
if (logs_pinned)
|
||||||
btrfs_log_new_name(trans, old_dentry, BTRFS_I(old_dir),
|
btrfs_log_new_name(trans, old_dentry, BTRFS_I(old_dir),
|
||||||
rename_ctx.index, new_dentry->d_parent);
|
rename_ctx.index, new_dentry->d_parent);
|
||||||
|
|
||||||
@@ -9205,6 +9248,10 @@ static int btrfs_rename(struct mnt_idmap *idmap,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
out_fail:
|
out_fail:
|
||||||
|
if (logs_pinned) {
|
||||||
|
btrfs_end_log_trans(root);
|
||||||
|
btrfs_end_log_trans(dest);
|
||||||
|
}
|
||||||
ret2 = btrfs_end_transaction(trans);
|
ret2 = btrfs_end_transaction(trans);
|
||||||
ret = ret ? ret : ret2;
|
ret = ret ? ret : ret2;
|
||||||
out_notrans:
|
out_notrans:
|
||||||
|
@@ -3174,6 +3174,12 @@ int btrfs_remove_chunk(struct btrfs_trans_handle *trans, u64 chunk_offset)
|
|||||||
device->bytes_used - dev_extent_len);
|
device->bytes_used - dev_extent_len);
|
||||||
atomic64_add(dev_extent_len, &fs_info->free_chunk_space);
|
atomic64_add(dev_extent_len, &fs_info->free_chunk_space);
|
||||||
btrfs_clear_space_info_full(fs_info);
|
btrfs_clear_space_info_full(fs_info);
|
||||||
|
|
||||||
|
if (list_empty(&device->post_commit_list)) {
|
||||||
|
list_add_tail(&device->post_commit_list,
|
||||||
|
&trans->transaction->dev_update_list);
|
||||||
|
}
|
||||||
|
|
||||||
mutex_unlock(&fs_info->chunk_mutex);
|
mutex_unlock(&fs_info->chunk_mutex);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -2513,7 +2513,7 @@ static int ceph_zero_objects(struct inode *inode, loff_t offset, loff_t length)
|
|||||||
s32 stripe_unit = ci->i_layout.stripe_unit;
|
s32 stripe_unit = ci->i_layout.stripe_unit;
|
||||||
s32 stripe_count = ci->i_layout.stripe_count;
|
s32 stripe_count = ci->i_layout.stripe_count;
|
||||||
s32 object_size = ci->i_layout.object_size;
|
s32 object_size = ci->i_layout.object_size;
|
||||||
u64 object_set_size = object_size * stripe_count;
|
u64 object_set_size = (u64) object_size * stripe_count;
|
||||||
u64 nearly, t;
|
u64 nearly, t;
|
||||||
|
|
||||||
/* round offset up to next period boundary */
|
/* round offset up to next period boundary */
|
||||||
|
@@ -1778,26 +1778,32 @@ static int f2fs_statfs_project(struct super_block *sb,
|
|||||||
|
|
||||||
limit = min_not_zero(dquot->dq_dqb.dqb_bsoftlimit,
|
limit = min_not_zero(dquot->dq_dqb.dqb_bsoftlimit,
|
||||||
dquot->dq_dqb.dqb_bhardlimit);
|
dquot->dq_dqb.dqb_bhardlimit);
|
||||||
if (limit)
|
|
||||||
limit >>= sb->s_blocksize_bits;
|
limit >>= sb->s_blocksize_bits;
|
||||||
|
|
||||||
if (limit && buf->f_blocks > limit) {
|
if (limit) {
|
||||||
|
uint64_t remaining = 0;
|
||||||
|
|
||||||
curblock = (dquot->dq_dqb.dqb_curspace +
|
curblock = (dquot->dq_dqb.dqb_curspace +
|
||||||
dquot->dq_dqb.dqb_rsvspace) >> sb->s_blocksize_bits;
|
dquot->dq_dqb.dqb_rsvspace) >> sb->s_blocksize_bits;
|
||||||
buf->f_blocks = limit;
|
if (limit > curblock)
|
||||||
buf->f_bfree = buf->f_bavail =
|
remaining = limit - curblock;
|
||||||
(buf->f_blocks > curblock) ?
|
|
||||||
(buf->f_blocks - curblock) : 0;
|
buf->f_blocks = min(buf->f_blocks, limit);
|
||||||
|
buf->f_bfree = min(buf->f_bfree, remaining);
|
||||||
|
buf->f_bavail = min(buf->f_bavail, remaining);
|
||||||
}
|
}
|
||||||
|
|
||||||
limit = min_not_zero(dquot->dq_dqb.dqb_isoftlimit,
|
limit = min_not_zero(dquot->dq_dqb.dqb_isoftlimit,
|
||||||
dquot->dq_dqb.dqb_ihardlimit);
|
dquot->dq_dqb.dqb_ihardlimit);
|
||||||
|
|
||||||
if (limit && buf->f_files > limit) {
|
if (limit) {
|
||||||
buf->f_files = limit;
|
uint64_t remaining = 0;
|
||||||
buf->f_ffree =
|
|
||||||
(buf->f_files > dquot->dq_dqb.dqb_curinodes) ?
|
if (limit > dquot->dq_dqb.dqb_curinodes)
|
||||||
(buf->f_files - dquot->dq_dqb.dqb_curinodes) : 0;
|
remaining = limit - dquot->dq_dqb.dqb_curinodes;
|
||||||
|
|
||||||
|
buf->f_files = min(buf->f_files, limit);
|
||||||
|
buf->f_ffree = min(buf->f_ffree, remaining);
|
||||||
}
|
}
|
||||||
|
|
||||||
spin_unlock(&dquot->dq_dqb_lock);
|
spin_unlock(&dquot->dq_dqb_lock);
|
||||||
|
@@ -2158,6 +2158,7 @@ int fuse_do_setattr(struct dentry *dentry, struct iattr *attr,
|
|||||||
int err;
|
int err;
|
||||||
bool trust_local_cmtime = is_wb;
|
bool trust_local_cmtime = is_wb;
|
||||||
bool fault_blocked = false;
|
bool fault_blocked = false;
|
||||||
|
u64 attr_version;
|
||||||
|
|
||||||
#ifdef CONFIG_FUSE_BPF
|
#ifdef CONFIG_FUSE_BPF
|
||||||
struct fuse_err_ret fer;
|
struct fuse_err_ret fer;
|
||||||
@@ -2252,6 +2253,8 @@ int fuse_do_setattr(struct dentry *dentry, struct iattr *attr,
|
|||||||
if (fc->handle_killpriv_v2 && !capable(CAP_FSETID))
|
if (fc->handle_killpriv_v2 && !capable(CAP_FSETID))
|
||||||
inarg.valid |= FATTR_KILL_SUIDGID;
|
inarg.valid |= FATTR_KILL_SUIDGID;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
attr_version = fuse_get_attr_version(fm->fc);
|
||||||
fuse_setattr_fill(fc, &args, inode, &inarg, &outarg);
|
fuse_setattr_fill(fc, &args, inode, &inarg, &outarg);
|
||||||
err = fuse_simple_request(fm, &args);
|
err = fuse_simple_request(fm, &args);
|
||||||
if (err) {
|
if (err) {
|
||||||
@@ -2277,6 +2280,14 @@ int fuse_do_setattr(struct dentry *dentry, struct iattr *attr,
|
|||||||
/* FIXME: clear I_DIRTY_SYNC? */
|
/* FIXME: clear I_DIRTY_SYNC? */
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (fi->attr_version > attr_version) {
|
||||||
|
/*
|
||||||
|
* Apply attributes, for example for fsnotify_change(), but set
|
||||||
|
* attribute timeout to zero.
|
||||||
|
*/
|
||||||
|
outarg.attr_valid = outarg.attr_valid_nsec = 0;
|
||||||
|
}
|
||||||
|
|
||||||
fuse_change_attributes_common(inode, &outarg.attr, NULL,
|
fuse_change_attributes_common(inode, &outarg.attr, NULL,
|
||||||
ATTR_TIMEOUT(&outarg),
|
ATTR_TIMEOUT(&outarg),
|
||||||
fuse_get_cache_mask(inode));
|
fuse_get_cache_mask(inode));
|
||||||
|
@@ -178,45 +178,30 @@ int dbMount(struct inode *ipbmap)
|
|||||||
dbmp_le = (struct dbmap_disk *) mp->data;
|
dbmp_le = (struct dbmap_disk *) mp->data;
|
||||||
bmp->db_mapsize = le64_to_cpu(dbmp_le->dn_mapsize);
|
bmp->db_mapsize = le64_to_cpu(dbmp_le->dn_mapsize);
|
||||||
bmp->db_nfree = le64_to_cpu(dbmp_le->dn_nfree);
|
bmp->db_nfree = le64_to_cpu(dbmp_le->dn_nfree);
|
||||||
|
|
||||||
bmp->db_l2nbperpage = le32_to_cpu(dbmp_le->dn_l2nbperpage);
|
bmp->db_l2nbperpage = le32_to_cpu(dbmp_le->dn_l2nbperpage);
|
||||||
if (bmp->db_l2nbperpage > L2PSIZE - L2MINBLOCKSIZE ||
|
|
||||||
bmp->db_l2nbperpage < 0) {
|
|
||||||
err = -EINVAL;
|
|
||||||
goto err_release_metapage;
|
|
||||||
}
|
|
||||||
|
|
||||||
bmp->db_numag = le32_to_cpu(dbmp_le->dn_numag);
|
bmp->db_numag = le32_to_cpu(dbmp_le->dn_numag);
|
||||||
if (!bmp->db_numag || bmp->db_numag > MAXAG) {
|
|
||||||
err = -EINVAL;
|
|
||||||
goto err_release_metapage;
|
|
||||||
}
|
|
||||||
|
|
||||||
bmp->db_maxlevel = le32_to_cpu(dbmp_le->dn_maxlevel);
|
bmp->db_maxlevel = le32_to_cpu(dbmp_le->dn_maxlevel);
|
||||||
bmp->db_maxag = le32_to_cpu(dbmp_le->dn_maxag);
|
bmp->db_maxag = le32_to_cpu(dbmp_le->dn_maxag);
|
||||||
bmp->db_agpref = le32_to_cpu(dbmp_le->dn_agpref);
|
bmp->db_agpref = le32_to_cpu(dbmp_le->dn_agpref);
|
||||||
if (bmp->db_maxag >= MAXAG || bmp->db_maxag < 0 ||
|
|
||||||
bmp->db_agpref >= MAXAG || bmp->db_agpref < 0) {
|
|
||||||
err = -EINVAL;
|
|
||||||
goto err_release_metapage;
|
|
||||||
}
|
|
||||||
|
|
||||||
bmp->db_aglevel = le32_to_cpu(dbmp_le->dn_aglevel);
|
bmp->db_aglevel = le32_to_cpu(dbmp_le->dn_aglevel);
|
||||||
bmp->db_agheight = le32_to_cpu(dbmp_le->dn_agheight);
|
bmp->db_agheight = le32_to_cpu(dbmp_le->dn_agheight);
|
||||||
bmp->db_agwidth = le32_to_cpu(dbmp_le->dn_agwidth);
|
bmp->db_agwidth = le32_to_cpu(dbmp_le->dn_agwidth);
|
||||||
if (!bmp->db_agwidth) {
|
|
||||||
err = -EINVAL;
|
|
||||||
goto err_release_metapage;
|
|
||||||
}
|
|
||||||
bmp->db_agstart = le32_to_cpu(dbmp_le->dn_agstart);
|
bmp->db_agstart = le32_to_cpu(dbmp_le->dn_agstart);
|
||||||
bmp->db_agl2size = le32_to_cpu(dbmp_le->dn_agl2size);
|
bmp->db_agl2size = le32_to_cpu(dbmp_le->dn_agl2size);
|
||||||
if (bmp->db_agl2size > L2MAXL2SIZE - L2MAXAG ||
|
|
||||||
bmp->db_agl2size < 0) {
|
|
||||||
err = -EINVAL;
|
|
||||||
goto err_release_metapage;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (((bmp->db_mapsize - 1) >> bmp->db_agl2size) > MAXAG) {
|
if ((bmp->db_l2nbperpage > L2PSIZE - L2MINBLOCKSIZE) ||
|
||||||
|
(bmp->db_l2nbperpage < 0) ||
|
||||||
|
!bmp->db_numag || (bmp->db_numag > MAXAG) ||
|
||||||
|
(bmp->db_maxag >= MAXAG) || (bmp->db_maxag < 0) ||
|
||||||
|
(bmp->db_agpref >= MAXAG) || (bmp->db_agpref < 0) ||
|
||||||
|
(bmp->db_agheight < 0) || (bmp->db_agheight > (L2LPERCTL >> 1)) ||
|
||||||
|
(bmp->db_agwidth < 1) || (bmp->db_agwidth > (LPERCTL / MAXAG)) ||
|
||||||
|
(bmp->db_agwidth > (1 << (L2LPERCTL - (bmp->db_agheight << 1)))) ||
|
||||||
|
(bmp->db_agstart < 0) ||
|
||||||
|
(bmp->db_agstart > (CTLTREESIZE - 1 - bmp->db_agwidth * (MAXAG - 1))) ||
|
||||||
|
(bmp->db_agl2size > L2MAXL2SIZE - L2MAXAG) ||
|
||||||
|
(bmp->db_agl2size < 0) ||
|
||||||
|
((bmp->db_mapsize - 1) >> bmp->db_agl2size) > MAXAG) {
|
||||||
err = -EINVAL;
|
err = -EINVAL;
|
||||||
goto err_release_metapage;
|
goto err_release_metapage;
|
||||||
}
|
}
|
||||||
|
@@ -2367,14 +2367,14 @@ static int attach_recursive_mnt(struct mount *source_mnt,
|
|||||||
hlist_for_each_entry_safe(child, n, &tree_list, mnt_hash) {
|
hlist_for_each_entry_safe(child, n, &tree_list, mnt_hash) {
|
||||||
struct mount *q;
|
struct mount *q;
|
||||||
hlist_del_init(&child->mnt_hash);
|
hlist_del_init(&child->mnt_hash);
|
||||||
q = __lookup_mnt(&child->mnt_parent->mnt,
|
|
||||||
child->mnt_mountpoint);
|
|
||||||
if (q)
|
|
||||||
mnt_change_mountpoint(child, smp, q);
|
|
||||||
/* Notice when we are propagating across user namespaces */
|
/* Notice when we are propagating across user namespaces */
|
||||||
if (child->mnt_parent->mnt_ns->user_ns != user_ns)
|
if (child->mnt_parent->mnt_ns->user_ns != user_ns)
|
||||||
lock_mnt_tree(child);
|
lock_mnt_tree(child);
|
||||||
child->mnt.mnt_flags &= ~MNT_LOCKED;
|
child->mnt.mnt_flags &= ~MNT_LOCKED;
|
||||||
|
q = __lookup_mnt(&child->mnt_parent->mnt,
|
||||||
|
child->mnt_mountpoint);
|
||||||
|
if (q)
|
||||||
|
mnt_change_mountpoint(child, smp, q);
|
||||||
commit_tree(child);
|
commit_tree(child);
|
||||||
}
|
}
|
||||||
put_mountpoint(smp);
|
put_mountpoint(smp);
|
||||||
|
@@ -555,6 +555,8 @@ nfs_fhget(struct super_block *sb, struct nfs_fh *fh, struct nfs_fattr *fattr)
|
|||||||
set_nlink(inode, fattr->nlink);
|
set_nlink(inode, fattr->nlink);
|
||||||
else if (fattr_supported & NFS_ATTR_FATTR_NLINK)
|
else if (fattr_supported & NFS_ATTR_FATTR_NLINK)
|
||||||
nfs_set_cache_invalid(inode, NFS_INO_INVALID_NLINK);
|
nfs_set_cache_invalid(inode, NFS_INO_INVALID_NLINK);
|
||||||
|
else
|
||||||
|
set_nlink(inode, 1);
|
||||||
if (fattr->valid & NFS_ATTR_FATTR_OWNER)
|
if (fattr->valid & NFS_ATTR_FATTR_OWNER)
|
||||||
inode->i_uid = fattr->uid;
|
inode->i_uid = fattr->uid;
|
||||||
else if (fattr_supported & NFS_ATTR_FATTR_OWNER)
|
else if (fattr_supported & NFS_ATTR_FATTR_OWNER)
|
||||||
|
@@ -6059,6 +6059,8 @@ static ssize_t nfs4_proc_get_acl(struct inode *inode, void *buf, size_t buflen,
|
|||||||
struct nfs_server *server = NFS_SERVER(inode);
|
struct nfs_server *server = NFS_SERVER(inode);
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
if (unlikely(NFS_FH(inode)->size == 0))
|
||||||
|
return -ENODATA;
|
||||||
if (!nfs4_server_supports_acls(server, type))
|
if (!nfs4_server_supports_acls(server, type))
|
||||||
return -EOPNOTSUPP;
|
return -EOPNOTSUPP;
|
||||||
ret = nfs_revalidate_inode(inode, NFS_INO_INVALID_CHANGE);
|
ret = nfs_revalidate_inode(inode, NFS_INO_INVALID_CHANGE);
|
||||||
@@ -6133,6 +6135,9 @@ static int nfs4_proc_set_acl(struct inode *inode, const void *buf,
|
|||||||
{
|
{
|
||||||
struct nfs4_exception exception = { };
|
struct nfs4_exception exception = { };
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
|
if (unlikely(NFS_FH(inode)->size == 0))
|
||||||
|
return -ENODATA;
|
||||||
do {
|
do {
|
||||||
err = __nfs4_proc_set_acl(inode, buf, buflen, type);
|
err = __nfs4_proc_set_acl(inode, buf, buflen, type);
|
||||||
trace_nfs4_set_acl(inode, err);
|
trace_nfs4_set_acl(inode, err);
|
||||||
@@ -10625,7 +10630,7 @@ const struct nfs4_minor_version_ops *nfs_v4_minor_ops[] = {
|
|||||||
|
|
||||||
static ssize_t nfs4_listxattr(struct dentry *dentry, char *list, size_t size)
|
static ssize_t nfs4_listxattr(struct dentry *dentry, char *list, size_t size)
|
||||||
{
|
{
|
||||||
ssize_t error, error2, error3;
|
ssize_t error, error2, error3, error4;
|
||||||
size_t left = size;
|
size_t left = size;
|
||||||
|
|
||||||
error = generic_listxattr(dentry, list, left);
|
error = generic_listxattr(dentry, list, left);
|
||||||
@@ -10648,8 +10653,16 @@ static ssize_t nfs4_listxattr(struct dentry *dentry, char *list, size_t size)
|
|||||||
error3 = nfs4_listxattr_nfs4_user(d_inode(dentry), list, left);
|
error3 = nfs4_listxattr_nfs4_user(d_inode(dentry), list, left);
|
||||||
if (error3 < 0)
|
if (error3 < 0)
|
||||||
return error3;
|
return error3;
|
||||||
|
if (list) {
|
||||||
|
list += error3;
|
||||||
|
left -= error3;
|
||||||
|
}
|
||||||
|
|
||||||
error += error2 + error3;
|
error4 = security_inode_listsecurity(d_inode(dentry), list, left);
|
||||||
|
if (error4 < 0)
|
||||||
|
return error4;
|
||||||
|
|
||||||
|
error += error2 + error3 + error4;
|
||||||
if (size && error > size)
|
if (size && error > size)
|
||||||
return -ERANGE;
|
return -ERANGE;
|
||||||
return error;
|
return error;
|
||||||
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user