Gentoo Archives: gentoo-commits

From: Mike Pagano <mpagano@g.o>
To: gentoo-commits@l.g.o
Subject: [gentoo-commits] proj/linux-patches:4.19 commit in: /
Date: Sat, 02 May 2020 19:24:08
Message-Id: 1588447424.d2975282a07e1079c8befdd23086852956fe4493.mpagano@gentoo
1 commit: d2975282a07e1079c8befdd23086852956fe4493
2 Author: Mike Pagano <mpagano <AT> gentoo <DOT> org>
3 AuthorDate: Sat May 2 19:23:44 2020 +0000
4 Commit: Mike Pagano <mpagano <AT> gentoo <DOT> org>
5 CommitDate: Sat May 2 19:23:44 2020 +0000
6 URL: https://gitweb.gentoo.org/proj/linux-patches.git/commit/?id=d2975282
7
8 Linux patch 4.19.120
9
10 Signed-off-by: Mike Pagano <mpagano <AT> gentoo.org>
11
12 0000_README | 4 +
13 1119_linux-4.19.120.patch | 1834 +++++++++++++++++++++++++++++++++++++++++++++
14 2 files changed, 1838 insertions(+)
15
16 diff --git a/0000_README b/0000_README
17 index 36e1c22..8850298 100644
18 --- a/0000_README
19 +++ b/0000_README
20 @@ -515,6 +515,10 @@ Patch: 1118_linux-4.19.119.patch
21 From: https://www.kernel.org
22 Desc: Linux 4.19.119
23
24 +Patch: 1119_linux-4.19.120.patch
25 +From: https://www.kernel.org
26 +Desc: Linux 4.19.120
27 +
28 Patch: 1500_XATTR_USER_PREFIX.patch
29 From: https://bugs.gentoo.org/show_bug.cgi?id=470644
30 Desc: Support for namespace user.pax.* on tmpfs.
31
32 diff --git a/1119_linux-4.19.120.patch b/1119_linux-4.19.120.patch
33 new file mode 100644
34 index 0000000..6f1d761
35 --- /dev/null
36 +++ b/1119_linux-4.19.120.patch
37 @@ -0,0 +1,1834 @@
38 +diff --git a/Makefile b/Makefile
39 +index 69c95fe6fba5..74b60bf2ff79 100644
40 +--- a/Makefile
41 ++++ b/Makefile
42 +@@ -1,7 +1,7 @@
43 + # SPDX-License-Identifier: GPL-2.0
44 + VERSION = 4
45 + PATCHLEVEL = 19
46 +-SUBLEVEL = 119
47 ++SUBLEVEL = 120
48 + EXTRAVERSION =
49 + NAME = "People's Front"
50 +
51 +diff --git a/arch/arm/boot/dts/bcm283x.dtsi b/arch/arm/boot/dts/bcm283x.dtsi
52 +index c9322a56300d..f9add515301f 100644
53 +--- a/arch/arm/boot/dts/bcm283x.dtsi
54 ++++ b/arch/arm/boot/dts/bcm283x.dtsi
55 +@@ -476,6 +476,7 @@
56 + "dsi0_ddr2",
57 + "dsi0_ddr";
58 +
59 ++ status = "disabled";
60 + };
61 +
62 + thermal: thermal@7e212000 {
63 +diff --git a/arch/arm64/include/asm/sysreg.h b/arch/arm64/include/asm/sysreg.h
64 +index 3091ae5975a3..ed99d941c462 100644
65 +--- a/arch/arm64/include/asm/sysreg.h
66 ++++ b/arch/arm64/include/asm/sysreg.h
67 +@@ -60,7 +60,9 @@
68 + #ifndef CONFIG_BROKEN_GAS_INST
69 +
70 + #ifdef __ASSEMBLY__
71 +-#define __emit_inst(x) .inst (x)
72 ++// The space separator is omitted so that __emit_inst(x) can be parsed as
73 ++// either an assembler directive or an assembler macro argument.
74 ++#define __emit_inst(x) .inst(x)
75 + #else
76 + #define __emit_inst(x) ".inst " __stringify((x)) "\n\t"
77 + #endif
78 +diff --git a/arch/x86/kernel/cpu/mshyperv.c b/arch/x86/kernel/cpu/mshyperv.c
79 +index fc93ae325515..f8b0fa2dbe37 100644
80 +--- a/arch/x86/kernel/cpu/mshyperv.c
81 ++++ b/arch/x86/kernel/cpu/mshyperv.c
82 +@@ -214,8 +214,8 @@ static void __init ms_hyperv_init_platform(void)
83 + ms_hyperv.misc_features = cpuid_edx(HYPERV_CPUID_FEATURES);
84 + ms_hyperv.hints = cpuid_eax(HYPERV_CPUID_ENLIGHTMENT_INFO);
85 +
86 +- pr_info("Hyper-V: features 0x%x, hints 0x%x\n",
87 +- ms_hyperv.features, ms_hyperv.hints);
88 ++ pr_info("Hyper-V: features 0x%x, hints 0x%x, misc 0x%x\n",
89 ++ ms_hyperv.features, ms_hyperv.hints, ms_hyperv.misc_features);
90 +
91 + ms_hyperv.max_vp_index = cpuid_eax(HYPERV_CPUID_IMPLEMENT_LIMITS);
92 + ms_hyperv.max_lp_index = cpuid_ebx(HYPERV_CPUID_IMPLEMENT_LIMITS);
93 +diff --git a/arch/x86/net/bpf_jit_comp.c b/arch/x86/net/bpf_jit_comp.c
94 +index a32fc3d99407..46ab92831251 100644
95 +--- a/arch/x86/net/bpf_jit_comp.c
96 ++++ b/arch/x86/net/bpf_jit_comp.c
97 +@@ -142,6 +142,19 @@ static bool is_ereg(u32 reg)
98 + BIT(BPF_REG_AX));
99 + }
100 +
101 ++/*
102 ++ * is_ereg_8l() == true if BPF register 'reg' is mapped to access x86-64
103 ++ * lower 8-bit registers dil,sil,bpl,spl,r8b..r15b, which need extra byte
104 ++ * of encoding. al,cl,dl,bl have simpler encoding.
105 ++ */
106 ++static bool is_ereg_8l(u32 reg)
107 ++{
108 ++ return is_ereg(reg) ||
109 ++ (1 << reg) & (BIT(BPF_REG_1) |
110 ++ BIT(BPF_REG_2) |
111 ++ BIT(BPF_REG_FP));
112 ++}
113 ++
114 + static bool is_axreg(u32 reg)
115 + {
116 + return reg == BPF_REG_0;
117 +@@ -751,9 +764,8 @@ st: if (is_imm8(insn->off))
118 + /* STX: *(u8*)(dst_reg + off) = src_reg */
119 + case BPF_STX | BPF_MEM | BPF_B:
120 + /* Emit 'mov byte ptr [rax + off], al' */
121 +- if (is_ereg(dst_reg) || is_ereg(src_reg) ||
122 +- /* We have to add extra byte for x86 SIL, DIL regs */
123 +- src_reg == BPF_REG_1 || src_reg == BPF_REG_2)
124 ++ if (is_ereg(dst_reg) || is_ereg_8l(src_reg))
125 ++ /* Add extra byte for eregs or SIL,DIL,BPL in src_reg */
126 + EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x88);
127 + else
128 + EMIT1(0x88);
129 +diff --git a/arch/x86/net/bpf_jit_comp32.c b/arch/x86/net/bpf_jit_comp32.c
130 +index 24d573bc550d..2eaf1900ba67 100644
131 +--- a/arch/x86/net/bpf_jit_comp32.c
132 ++++ b/arch/x86/net/bpf_jit_comp32.c
133 +@@ -1830,7 +1830,9 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
134 + STACK_VAR(dst_hi));
135 + EMIT(0x0, 4);
136 + } else {
137 +- EMIT3(0xC7, add_1reg(0xC0, dst_hi), 0);
138 ++ /* xor dst_hi,dst_hi */
139 ++ EMIT2(0x33,
140 ++ add_2reg(0xC0, dst_hi, dst_hi));
141 + }
142 + break;
143 + case BPF_DW:
144 +@@ -1968,8 +1970,8 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
145 + goto emit_cond_jmp_signed;
146 + }
147 + case BPF_JMP | BPF_JSET | BPF_X: {
148 +- u8 dreg_lo = dstk ? IA32_EAX : dst_lo;
149 +- u8 dreg_hi = dstk ? IA32_EDX : dst_hi;
150 ++ u8 dreg_lo = IA32_EAX;
151 ++ u8 dreg_hi = IA32_EDX;
152 + u8 sreg_lo = sstk ? IA32_ECX : src_lo;
153 + u8 sreg_hi = sstk ? IA32_EBX : src_hi;
154 +
155 +@@ -1978,6 +1980,12 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
156 + STACK_VAR(dst_lo));
157 + EMIT3(0x8B, add_2reg(0x40, IA32_EBP, IA32_EDX),
158 + STACK_VAR(dst_hi));
159 ++ } else {
160 ++ /* mov dreg_lo,dst_lo */
161 ++ EMIT2(0x89, add_2reg(0xC0, dreg_lo, dst_lo));
162 ++ /* mov dreg_hi,dst_hi */
163 ++ EMIT2(0x89,
164 ++ add_2reg(0xC0, dreg_hi, dst_hi));
165 + }
166 +
167 + if (sstk) {
168 +@@ -1996,8 +2004,8 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
169 + }
170 + case BPF_JMP | BPF_JSET | BPF_K: {
171 + u32 hi;
172 +- u8 dreg_lo = dstk ? IA32_EAX : dst_lo;
173 +- u8 dreg_hi = dstk ? IA32_EDX : dst_hi;
174 ++ u8 dreg_lo = IA32_EAX;
175 ++ u8 dreg_hi = IA32_EDX;
176 + u8 sreg_lo = IA32_ECX;
177 + u8 sreg_hi = IA32_EBX;
178 +
179 +@@ -2006,6 +2014,12 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image,
180 + STACK_VAR(dst_lo));
181 + EMIT3(0x8B, add_2reg(0x40, IA32_EBP, IA32_EDX),
182 + STACK_VAR(dst_hi));
183 ++ } else {
184 ++ /* mov dreg_lo,dst_lo */
185 ++ EMIT2(0x89, add_2reg(0xC0, dreg_lo, dst_lo));
186 ++ /* mov dreg_hi,dst_hi */
187 ++ EMIT2(0x89,
188 ++ add_2reg(0xC0, dreg_hi, dst_hi));
189 + }
190 + hi = imm32 & (1<<31) ? (u32)~0 : 0;
191 +
192 +diff --git a/drivers/android/binder_alloc.c b/drivers/android/binder_alloc.c
193 +index c3fdd79c4082..6e5e6fb7e4c9 100644
194 +--- a/drivers/android/binder_alloc.c
195 ++++ b/drivers/android/binder_alloc.c
196 +@@ -970,8 +970,8 @@ enum lru_status binder_alloc_free_page(struct list_head *item,
197 + mm = alloc->vma_vm_mm;
198 + if (!mmget_not_zero(mm))
199 + goto err_mmget;
200 +- if (!down_write_trylock(&mm->mmap_sem))
201 +- goto err_down_write_mmap_sem_failed;
202 ++ if (!down_read_trylock(&mm->mmap_sem))
203 ++ goto err_down_read_mmap_sem_failed;
204 + vma = binder_alloc_get_vma(alloc);
205 +
206 + list_lru_isolate(lru, item);
207 +@@ -986,7 +986,7 @@ enum lru_status binder_alloc_free_page(struct list_head *item,
208 +
209 + trace_binder_unmap_user_end(alloc, index);
210 + }
211 +- up_write(&mm->mmap_sem);
212 ++ up_read(&mm->mmap_sem);
213 + mmput(mm);
214 +
215 + trace_binder_unmap_kernel_start(alloc, index);
216 +@@ -1001,7 +1001,7 @@ enum lru_status binder_alloc_free_page(struct list_head *item,
217 + mutex_unlock(&alloc->mutex);
218 + return LRU_REMOVED_RETRY;
219 +
220 +-err_down_write_mmap_sem_failed:
221 ++err_down_read_mmap_sem_failed:
222 + mmput_async(mm);
223 + err_mmget:
224 + err_page_already_freed:
225 +diff --git a/drivers/hwmon/jc42.c b/drivers/hwmon/jc42.c
226 +index e5234f953a6d..b6e5aaa54963 100644
227 +--- a/drivers/hwmon/jc42.c
228 ++++ b/drivers/hwmon/jc42.c
229 +@@ -527,7 +527,7 @@ static int jc42_probe(struct i2c_client *client, const struct i2c_device_id *id)
230 + }
231 + data->config = config;
232 +
233 +- hwmon_dev = devm_hwmon_device_register_with_info(dev, client->name,
234 ++ hwmon_dev = devm_hwmon_device_register_with_info(dev, "jc42",
235 + data, &jc42_chip_info,
236 + NULL);
237 + return PTR_ERR_OR_ZERO(hwmon_dev);
238 +diff --git a/drivers/i2c/busses/i2c-altera.c b/drivers/i2c/busses/i2c-altera.c
239 +index a1cdcfc74acf..8915ee30a5b4 100644
240 +--- a/drivers/i2c/busses/i2c-altera.c
241 ++++ b/drivers/i2c/busses/i2c-altera.c
242 +@@ -395,7 +395,6 @@ static int altr_i2c_probe(struct platform_device *pdev)
243 + struct altr_i2c_dev *idev = NULL;
244 + struct resource *res;
245 + int irq, ret;
246 +- u32 val;
247 +
248 + idev = devm_kzalloc(&pdev->dev, sizeof(*idev), GFP_KERNEL);
249 + if (!idev)
250 +@@ -422,17 +421,17 @@ static int altr_i2c_probe(struct platform_device *pdev)
251 + init_completion(&idev->msg_complete);
252 + spin_lock_init(&idev->lock);
253 +
254 +- val = device_property_read_u32(idev->dev, "fifo-size",
255 ++ ret = device_property_read_u32(idev->dev, "fifo-size",
256 + &idev->fifo_size);
257 +- if (val) {
258 ++ if (ret) {
259 + dev_err(&pdev->dev, "FIFO size set to default of %d\n",
260 + ALTR_I2C_DFLT_FIFO_SZ);
261 + idev->fifo_size = ALTR_I2C_DFLT_FIFO_SZ;
262 + }
263 +
264 +- val = device_property_read_u32(idev->dev, "clock-frequency",
265 ++ ret = device_property_read_u32(idev->dev, "clock-frequency",
266 + &idev->bus_clk_rate);
267 +- if (val) {
268 ++ if (ret) {
269 + dev_err(&pdev->dev, "Default to 100kHz\n");
270 + idev->bus_clk_rate = 100000; /* default clock rate */
271 + }
272 +diff --git a/drivers/iio/adc/ad7793.c b/drivers/iio/adc/ad7793.c
273 +index d4bbe5b53318..23a6e7baa396 100644
274 +--- a/drivers/iio/adc/ad7793.c
275 ++++ b/drivers/iio/adc/ad7793.c
276 +@@ -542,7 +542,7 @@ static const struct iio_info ad7797_info = {
277 + .read_raw = &ad7793_read_raw,
278 + .write_raw = &ad7793_write_raw,
279 + .write_raw_get_fmt = &ad7793_write_raw_get_fmt,
280 +- .attrs = &ad7793_attribute_group,
281 ++ .attrs = &ad7797_attribute_group,
282 + .validate_trigger = ad_sd_validate_trigger,
283 + };
284 +
285 +diff --git a/drivers/mtd/chips/cfi_cmdset_0002.c b/drivers/mtd/chips/cfi_cmdset_0002.c
286 +index ba44ea6d497e..1dbc9554a078 100644
287 +--- a/drivers/mtd/chips/cfi_cmdset_0002.c
288 ++++ b/drivers/mtd/chips/cfi_cmdset_0002.c
289 +@@ -1882,7 +1882,11 @@ static int __xipram do_write_buffer(struct map_info *map, struct flchip *chip,
290 + continue;
291 + }
292 +
293 +- if (time_after(jiffies, timeo) && !chip_ready(map, adr))
294 ++ /*
295 ++ * We check "time_after" and "!chip_good" before checking "chip_good" to avoid
296 ++ * the failure due to scheduling.
297 ++ */
298 ++ if (time_after(jiffies, timeo) && !chip_good(map, adr, datum))
299 + break;
300 +
301 + if (chip_good(map, adr, datum)) {
302 +diff --git a/drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b/drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
303 +index 8350c0c9b89d..5934ec9b6a31 100644
304 +--- a/drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
305 ++++ b/drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
306 +@@ -3748,7 +3748,7 @@ int t4_phy_fw_ver(struct adapter *adap, int *phy_fw_ver)
307 + FW_PARAMS_PARAM_Z_V(FW_PARAMS_PARAM_DEV_PHYFW_VERSION));
308 + ret = t4_query_params(adap, adap->mbox, adap->pf, 0, 1,
309 + &param, &val);
310 +- if (ret < 0)
311 ++ if (ret)
312 + return ret;
313 + *phy_fw_ver = val;
314 + return 0;
315 +diff --git a/drivers/net/ethernet/freescale/fec.h b/drivers/net/ethernet/freescale/fec.h
316 +index bf80855dd0dd..d06a89e99872 100644
317 +--- a/drivers/net/ethernet/freescale/fec.h
318 ++++ b/drivers/net/ethernet/freescale/fec.h
319 +@@ -488,6 +488,12 @@ struct fec_enet_priv_rx_q {
320 + struct sk_buff *rx_skbuff[RX_RING_SIZE];
321 + };
322 +
323 ++struct fec_stop_mode_gpr {
324 ++ struct regmap *gpr;
325 ++ u8 reg;
326 ++ u8 bit;
327 ++};
328 ++
329 + /* The FEC buffer descriptors track the ring buffers. The rx_bd_base and
330 + * tx_bd_base always point to the base of the buffer descriptors. The
331 + * cur_rx and cur_tx point to the currently available buffer.
332 +@@ -563,6 +569,7 @@ struct fec_enet_private {
333 + int hwts_tx_en;
334 + struct delayed_work time_keep;
335 + struct regulator *reg_phy;
336 ++ struct fec_stop_mode_gpr stop_gpr;
337 +
338 + unsigned int tx_align;
339 + unsigned int rx_align;
340 +diff --git a/drivers/net/ethernet/freescale/fec_main.c b/drivers/net/ethernet/freescale/fec_main.c
341 +index 9142992ccd5a..48c58f93b124 100644
342 +--- a/drivers/net/ethernet/freescale/fec_main.c
343 ++++ b/drivers/net/ethernet/freescale/fec_main.c
344 +@@ -62,6 +62,8 @@
345 + #include <linux/if_vlan.h>
346 + #include <linux/pinctrl/consumer.h>
347 + #include <linux/prefetch.h>
348 ++#include <linux/mfd/syscon.h>
349 ++#include <linux/regmap.h>
350 + #include <soc/imx/cpuidle.h>
351 +
352 + #include <asm/cacheflush.h>
353 +@@ -84,6 +86,56 @@ static void fec_enet_itr_coal_init(struct net_device *ndev);
354 + #define FEC_ENET_OPD_V 0xFFF0
355 + #define FEC_MDIO_PM_TIMEOUT 100 /* ms */
356 +
357 ++struct fec_devinfo {
358 ++ u32 quirks;
359 ++ u8 stop_gpr_reg;
360 ++ u8 stop_gpr_bit;
361 ++};
362 ++
363 ++static const struct fec_devinfo fec_imx25_info = {
364 ++ .quirks = FEC_QUIRK_USE_GASKET | FEC_QUIRK_MIB_CLEAR |
365 ++ FEC_QUIRK_HAS_FRREG,
366 ++};
367 ++
368 ++static const struct fec_devinfo fec_imx27_info = {
369 ++ .quirks = FEC_QUIRK_MIB_CLEAR | FEC_QUIRK_HAS_FRREG,
370 ++};
371 ++
372 ++static const struct fec_devinfo fec_imx28_info = {
373 ++ .quirks = FEC_QUIRK_ENET_MAC | FEC_QUIRK_SWAP_FRAME |
374 ++ FEC_QUIRK_SINGLE_MDIO | FEC_QUIRK_HAS_RACC |
375 ++ FEC_QUIRK_HAS_FRREG,
376 ++};
377 ++
378 ++static const struct fec_devinfo fec_imx6q_info = {
379 ++ .quirks = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
380 ++ FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
381 ++ FEC_QUIRK_HAS_VLAN | FEC_QUIRK_ERR006358 |
382 ++ FEC_QUIRK_HAS_RACC,
383 ++ .stop_gpr_reg = 0x34,
384 ++ .stop_gpr_bit = 27,
385 ++};
386 ++
387 ++static const struct fec_devinfo fec_mvf600_info = {
388 ++ .quirks = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_RACC,
389 ++};
390 ++
391 ++static const struct fec_devinfo fec_imx6x_info = {
392 ++ .quirks = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
393 ++ FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
394 ++ FEC_QUIRK_HAS_VLAN | FEC_QUIRK_HAS_AVB |
395 ++ FEC_QUIRK_ERR007885 | FEC_QUIRK_BUG_CAPTURE |
396 ++ FEC_QUIRK_HAS_RACC | FEC_QUIRK_HAS_COALESCE,
397 ++};
398 ++
399 ++static const struct fec_devinfo fec_imx6ul_info = {
400 ++ .quirks = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
401 ++ FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
402 ++ FEC_QUIRK_HAS_VLAN | FEC_QUIRK_ERR007885 |
403 ++ FEC_QUIRK_BUG_CAPTURE | FEC_QUIRK_HAS_RACC |
404 ++ FEC_QUIRK_HAS_COALESCE,
405 ++};
406 ++
407 + static struct platform_device_id fec_devtype[] = {
408 + {
409 + /* keep it for coldfire */
410 +@@ -91,39 +143,25 @@ static struct platform_device_id fec_devtype[] = {
411 + .driver_data = 0,
412 + }, {
413 + .name = "imx25-fec",
414 +- .driver_data = FEC_QUIRK_USE_GASKET | FEC_QUIRK_MIB_CLEAR |
415 +- FEC_QUIRK_HAS_FRREG,
416 ++ .driver_data = (kernel_ulong_t)&fec_imx25_info,
417 + }, {
418 + .name = "imx27-fec",
419 +- .driver_data = FEC_QUIRK_MIB_CLEAR | FEC_QUIRK_HAS_FRREG,
420 ++ .driver_data = (kernel_ulong_t)&fec_imx27_info,
421 + }, {
422 + .name = "imx28-fec",
423 +- .driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_SWAP_FRAME |
424 +- FEC_QUIRK_SINGLE_MDIO | FEC_QUIRK_HAS_RACC |
425 +- FEC_QUIRK_HAS_FRREG,
426 ++ .driver_data = (kernel_ulong_t)&fec_imx28_info,
427 + }, {
428 + .name = "imx6q-fec",
429 +- .driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
430 +- FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
431 +- FEC_QUIRK_HAS_VLAN | FEC_QUIRK_ERR006358 |
432 +- FEC_QUIRK_HAS_RACC,
433 ++ .driver_data = (kernel_ulong_t)&fec_imx6q_info,
434 + }, {
435 + .name = "mvf600-fec",
436 +- .driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_RACC,
437 ++ .driver_data = (kernel_ulong_t)&fec_mvf600_info,
438 + }, {
439 + .name = "imx6sx-fec",
440 +- .driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
441 +- FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
442 +- FEC_QUIRK_HAS_VLAN | FEC_QUIRK_HAS_AVB |
443 +- FEC_QUIRK_ERR007885 | FEC_QUIRK_BUG_CAPTURE |
444 +- FEC_QUIRK_HAS_RACC | FEC_QUIRK_HAS_COALESCE,
445 ++ .driver_data = (kernel_ulong_t)&fec_imx6x_info,
446 + }, {
447 + .name = "imx6ul-fec",
448 +- .driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
449 +- FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
450 +- FEC_QUIRK_HAS_VLAN | FEC_QUIRK_ERR007885 |
451 +- FEC_QUIRK_BUG_CAPTURE | FEC_QUIRK_HAS_RACC |
452 +- FEC_QUIRK_HAS_COALESCE,
453 ++ .driver_data = (kernel_ulong_t)&fec_imx6ul_info,
454 + }, {
455 + /* sentinel */
456 + }
457 +@@ -1089,11 +1127,28 @@ fec_restart(struct net_device *ndev)
458 +
459 + }
460 +
461 ++static void fec_enet_stop_mode(struct fec_enet_private *fep, bool enabled)
462 ++{
463 ++ struct fec_platform_data *pdata = fep->pdev->dev.platform_data;
464 ++ struct fec_stop_mode_gpr *stop_gpr = &fep->stop_gpr;
465 ++
466 ++ if (stop_gpr->gpr) {
467 ++ if (enabled)
468 ++ regmap_update_bits(stop_gpr->gpr, stop_gpr->reg,
469 ++ BIT(stop_gpr->bit),
470 ++ BIT(stop_gpr->bit));
471 ++ else
472 ++ regmap_update_bits(stop_gpr->gpr, stop_gpr->reg,
473 ++ BIT(stop_gpr->bit), 0);
474 ++ } else if (pdata && pdata->sleep_mode_enable) {
475 ++ pdata->sleep_mode_enable(enabled);
476 ++ }
477 ++}
478 ++
479 + static void
480 + fec_stop(struct net_device *ndev)
481 + {
482 + struct fec_enet_private *fep = netdev_priv(ndev);
483 +- struct fec_platform_data *pdata = fep->pdev->dev.platform_data;
484 + u32 rmii_mode = readl(fep->hwp + FEC_R_CNTRL) & (1 << 8);
485 + u32 val;
486 +
487 +@@ -1122,9 +1177,7 @@ fec_stop(struct net_device *ndev)
488 + val = readl(fep->hwp + FEC_ECNTRL);
489 + val |= (FEC_ECR_MAGICEN | FEC_ECR_SLEEP);
490 + writel(val, fep->hwp + FEC_ECNTRL);
491 +-
492 +- if (pdata && pdata->sleep_mode_enable)
493 +- pdata->sleep_mode_enable(true);
494 ++ fec_enet_stop_mode(fep, true);
495 + }
496 + writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);
497 +
498 +@@ -3347,6 +3400,37 @@ static int fec_enet_get_irq_cnt(struct platform_device *pdev)
499 + return irq_cnt;
500 + }
501 +
502 ++static int fec_enet_init_stop_mode(struct fec_enet_private *fep,
503 ++ struct fec_devinfo *dev_info,
504 ++ struct device_node *np)
505 ++{
506 ++ struct device_node *gpr_np;
507 ++ int ret = 0;
508 ++
509 ++ if (!dev_info)
510 ++ return 0;
511 ++
512 ++ gpr_np = of_parse_phandle(np, "gpr", 0);
513 ++ if (!gpr_np)
514 ++ return 0;
515 ++
516 ++ fep->stop_gpr.gpr = syscon_node_to_regmap(gpr_np);
517 ++ if (IS_ERR(fep->stop_gpr.gpr)) {
518 ++ dev_err(&fep->pdev->dev, "could not find gpr regmap\n");
519 ++ ret = PTR_ERR(fep->stop_gpr.gpr);
520 ++ fep->stop_gpr.gpr = NULL;
521 ++ goto out;
522 ++ }
523 ++
524 ++ fep->stop_gpr.reg = dev_info->stop_gpr_reg;
525 ++ fep->stop_gpr.bit = dev_info->stop_gpr_bit;
526 ++
527 ++out:
528 ++ of_node_put(gpr_np);
529 ++
530 ++ return ret;
531 ++}
532 ++
533 + static int
534 + fec_probe(struct platform_device *pdev)
535 + {
536 +@@ -3362,6 +3446,7 @@ fec_probe(struct platform_device *pdev)
537 + int num_rx_qs;
538 + char irq_name[8];
539 + int irq_cnt;
540 ++ struct fec_devinfo *dev_info;
541 +
542 + fec_enet_get_queue_num(pdev, &num_tx_qs, &num_rx_qs);
543 +
544 +@@ -3379,7 +3464,9 @@ fec_probe(struct platform_device *pdev)
545 + of_id = of_match_device(fec_dt_ids, &pdev->dev);
546 + if (of_id)
547 + pdev->id_entry = of_id->data;
548 +- fep->quirks = pdev->id_entry->driver_data;
549 ++ dev_info = (struct fec_devinfo *)pdev->id_entry->driver_data;
550 ++ if (dev_info)
551 ++ fep->quirks = dev_info->quirks;
552 +
553 + fep->netdev = ndev;
554 + fep->num_rx_queues = num_rx_qs;
555 +@@ -3414,6 +3501,10 @@ fec_probe(struct platform_device *pdev)
556 + if (of_get_property(np, "fsl,magic-packet", NULL))
557 + fep->wol_flag |= FEC_WOL_HAS_MAGIC_PACKET;
558 +
559 ++ ret = fec_enet_init_stop_mode(fep, dev_info, np);
560 ++ if (ret)
561 ++ goto failed_stop_mode;
562 ++
563 + phy_node = of_parse_phandle(np, "phy-handle", 0);
564 + if (!phy_node && of_phy_is_fixed_link(np)) {
565 + ret = of_phy_register_fixed_link(np);
566 +@@ -3583,6 +3674,7 @@ failed_clk:
567 + if (of_phy_is_fixed_link(np))
568 + of_phy_deregister_fixed_link(np);
569 + of_node_put(phy_node);
570 ++failed_stop_mode:
571 + failed_phy:
572 + dev_id--;
573 + failed_ioremap:
574 +@@ -3660,7 +3752,6 @@ static int __maybe_unused fec_resume(struct device *dev)
575 + {
576 + struct net_device *ndev = dev_get_drvdata(dev);
577 + struct fec_enet_private *fep = netdev_priv(ndev);
578 +- struct fec_platform_data *pdata = fep->pdev->dev.platform_data;
579 + int ret;
580 + int val;
581 +
582 +@@ -3678,8 +3769,8 @@ static int __maybe_unused fec_resume(struct device *dev)
583 + goto failed_clk;
584 + }
585 + if (fep->wol_flag & FEC_WOL_FLAG_ENABLE) {
586 +- if (pdata && pdata->sleep_mode_enable)
587 +- pdata->sleep_mode_enable(false);
588 ++ fec_enet_stop_mode(fep, false);
589 ++
590 + val = readl(fep->hwp + FEC_ECNTRL);
591 + val &= ~(FEC_ECR_MAGICEN | FEC_ECR_SLEEP);
592 + writel(val, fep->hwp + FEC_ECNTRL);
593 +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c b/drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
594 +index d4ec93bde4de..2266c09b741a 100644
595 +--- a/drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
596 ++++ b/drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c
597 +@@ -796,7 +796,7 @@ struct mlx5_fw_tracer *mlx5_fw_tracer_create(struct mlx5_core_dev *dev)
598 + return NULL;
599 + }
600 +
601 +- tracer = kzalloc(sizeof(*tracer), GFP_KERNEL);
602 ++ tracer = kvzalloc(sizeof(*tracer), GFP_KERNEL);
603 + if (!tracer)
604 + return ERR_PTR(-ENOMEM);
605 +
606 +@@ -842,7 +842,7 @@ destroy_workqueue:
607 + tracer->dev = NULL;
608 + destroy_workqueue(tracer->work_queue);
609 + free_tracer:
610 +- kfree(tracer);
611 ++ kvfree(tracer);
612 + return ERR_PTR(err);
613 + }
614 +
615 +@@ -919,7 +919,7 @@ void mlx5_fw_tracer_destroy(struct mlx5_fw_tracer *tracer)
616 + mlx5_fw_tracer_destroy_log_buf(tracer);
617 + flush_workqueue(tracer->work_queue);
618 + destroy_workqueue(tracer->work_queue);
619 +- kfree(tracer);
620 ++ kvfree(tracer);
621 + }
622 +
623 + void mlx5_fw_tracer_event(struct mlx5_core_dev *dev, struct mlx5_eqe *eqe)
624 +diff --git a/drivers/net/ethernet/qlogic/qed/qed_dev.c b/drivers/net/ethernet/qlogic/qed/qed_dev.c
625 +index a6a9688db307..e50fc8f714dc 100644
626 +--- a/drivers/net/ethernet/qlogic/qed/qed_dev.c
627 ++++ b/drivers/net/ethernet/qlogic/qed/qed_dev.c
628 +@@ -3316,26 +3316,20 @@ static void qed_chain_free_single(struct qed_dev *cdev,
629 +
630 + static void qed_chain_free_pbl(struct qed_dev *cdev, struct qed_chain *p_chain)
631 + {
632 +- void **pp_virt_addr_tbl = p_chain->pbl.pp_virt_addr_tbl;
633 ++ struct addr_tbl_entry *pp_addr_tbl = p_chain->pbl.pp_addr_tbl;
634 + u32 page_cnt = p_chain->page_cnt, i, pbl_size;
635 +- u8 *p_pbl_virt = p_chain->pbl_sp.p_virt_table;
636 +
637 +- if (!pp_virt_addr_tbl)
638 ++ if (!pp_addr_tbl)
639 + return;
640 +
641 +- if (!p_pbl_virt)
642 +- goto out;
643 +-
644 + for (i = 0; i < page_cnt; i++) {
645 +- if (!pp_virt_addr_tbl[i])
646 ++ if (!pp_addr_tbl[i].virt_addr || !pp_addr_tbl[i].dma_map)
647 + break;
648 +
649 + dma_free_coherent(&cdev->pdev->dev,
650 + QED_CHAIN_PAGE_SIZE,
651 +- pp_virt_addr_tbl[i],
652 +- *(dma_addr_t *)p_pbl_virt);
653 +-
654 +- p_pbl_virt += QED_CHAIN_PBL_ENTRY_SIZE;
655 ++ pp_addr_tbl[i].virt_addr,
656 ++ pp_addr_tbl[i].dma_map);
657 + }
658 +
659 + pbl_size = page_cnt * QED_CHAIN_PBL_ENTRY_SIZE;
660 +@@ -3345,9 +3339,9 @@ static void qed_chain_free_pbl(struct qed_dev *cdev, struct qed_chain *p_chain)
661 + pbl_size,
662 + p_chain->pbl_sp.p_virt_table,
663 + p_chain->pbl_sp.p_phys_table);
664 +-out:
665 +- vfree(p_chain->pbl.pp_virt_addr_tbl);
666 +- p_chain->pbl.pp_virt_addr_tbl = NULL;
667 ++
668 ++ vfree(p_chain->pbl.pp_addr_tbl);
669 ++ p_chain->pbl.pp_addr_tbl = NULL;
670 + }
671 +
672 + void qed_chain_free(struct qed_dev *cdev, struct qed_chain *p_chain)
673 +@@ -3448,19 +3442,19 @@ qed_chain_alloc_pbl(struct qed_dev *cdev,
674 + {
675 + u32 page_cnt = p_chain->page_cnt, size, i;
676 + dma_addr_t p_phys = 0, p_pbl_phys = 0;
677 +- void **pp_virt_addr_tbl = NULL;
678 ++ struct addr_tbl_entry *pp_addr_tbl;
679 + u8 *p_pbl_virt = NULL;
680 + void *p_virt = NULL;
681 +
682 +- size = page_cnt * sizeof(*pp_virt_addr_tbl);
683 +- pp_virt_addr_tbl = vzalloc(size);
684 +- if (!pp_virt_addr_tbl)
685 ++ size = page_cnt * sizeof(*pp_addr_tbl);
686 ++ pp_addr_tbl = vzalloc(size);
687 ++ if (!pp_addr_tbl)
688 + return -ENOMEM;
689 +
690 + /* The allocation of the PBL table is done with its full size, since it
691 + * is expected to be successive.
692 + * qed_chain_init_pbl_mem() is called even in a case of an allocation
693 +- * failure, since pp_virt_addr_tbl was previously allocated, and it
694 ++ * failure, since tbl was previously allocated, and it
695 + * should be saved to allow its freeing during the error flow.
696 + */
697 + size = page_cnt * QED_CHAIN_PBL_ENTRY_SIZE;
698 +@@ -3474,8 +3468,7 @@ qed_chain_alloc_pbl(struct qed_dev *cdev,
699 + p_chain->b_external_pbl = true;
700 + }
701 +
702 +- qed_chain_init_pbl_mem(p_chain, p_pbl_virt, p_pbl_phys,
703 +- pp_virt_addr_tbl);
704 ++ qed_chain_init_pbl_mem(p_chain, p_pbl_virt, p_pbl_phys, pp_addr_tbl);
705 + if (!p_pbl_virt)
706 + return -ENOMEM;
707 +
708 +@@ -3494,7 +3487,8 @@ qed_chain_alloc_pbl(struct qed_dev *cdev,
709 + /* Fill the PBL table with the physical address of the page */
710 + *(dma_addr_t *)p_pbl_virt = p_phys;
711 + /* Keep the virtual address of the page */
712 +- p_chain->pbl.pp_virt_addr_tbl[i] = p_virt;
713 ++ p_chain->pbl.pp_addr_tbl[i].virt_addr = p_virt;
714 ++ p_chain->pbl.pp_addr_tbl[i].dma_map = p_phys;
715 +
716 + p_pbl_virt += QED_CHAIN_PBL_ENTRY_SIZE;
717 + }
718 +diff --git a/drivers/pci/quirks.c b/drivers/pci/quirks.c
719 +index 9e20ace30b62..ca41cff2e68c 100644
720 +--- a/drivers/pci/quirks.c
721 ++++ b/drivers/pci/quirks.c
722 +@@ -5282,3 +5282,21 @@ out_disable:
723 + DECLARE_PCI_FIXUP_CLASS_FINAL(PCI_VENDOR_ID_NVIDIA, 0x13b1,
724 + PCI_CLASS_DISPLAY_VGA, 8,
725 + quirk_reset_lenovo_thinkpad_p50_nvgpu);
726 ++
727 ++/*
728 ++ * Device [1b21:2142]
729 ++ * When in D0, PME# doesn't get asserted when plugging USB 3.0 device.
730 ++ */
731 ++static void pci_fixup_no_d0_pme(struct pci_dev *dev)
732 ++{
733 ++ pci_info(dev, "PME# does not work under D0, disabling it\n");
734 ++ dev->pme_support &= ~(PCI_PM_CAP_PME_D0 >> PCI_PM_CAP_PME_SHIFT);
735 ++}
736 ++DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_ASMEDIA, 0x2142, pci_fixup_no_d0_pme);
737 ++
738 ++static void apex_pci_fixup_class(struct pci_dev *pdev)
739 ++{
740 ++ pdev->class = (PCI_CLASS_SYSTEM_OTHER << 8) | pdev->class;
741 ++}
742 ++DECLARE_PCI_FIXUP_CLASS_HEADER(0x1ac1, 0x089a,
743 ++ PCI_CLASS_NOT_DEFINED, 8, apex_pci_fixup_class);
744 +diff --git a/drivers/remoteproc/remoteproc_core.c b/drivers/remoteproc/remoteproc_core.c
745 +index abbef17c97ee..d5ff272fde34 100644
746 +--- a/drivers/remoteproc/remoteproc_core.c
747 ++++ b/drivers/remoteproc/remoteproc_core.c
748 +@@ -289,7 +289,7 @@ void rproc_free_vring(struct rproc_vring *rvring)
749 + {
750 + int size = PAGE_ALIGN(vring_size(rvring->len, rvring->align));
751 + struct rproc *rproc = rvring->rvdev->rproc;
752 +- int idx = rvring->rvdev->vring - rvring;
753 ++ int idx = rvring - rvring->rvdev->vring;
754 + struct fw_rsc_vdev *rsc;
755 +
756 + dma_free_coherent(rproc->dev.parent, size, rvring->va, rvring->dma);
757 +diff --git a/drivers/staging/gasket/apex_driver.c b/drivers/staging/gasket/apex_driver.c
758 +index 0cef1d6d2e2b..99da735bdf6d 100644
759 +--- a/drivers/staging/gasket/apex_driver.c
760 ++++ b/drivers/staging/gasket/apex_driver.c
761 +@@ -578,13 +578,6 @@ static const struct pci_device_id apex_pci_ids[] = {
762 + { PCI_DEVICE(APEX_PCI_VENDOR_ID, APEX_PCI_DEVICE_ID) }, { 0 }
763 + };
764 +
765 +-static void apex_pci_fixup_class(struct pci_dev *pdev)
766 +-{
767 +- pdev->class = (PCI_CLASS_SYSTEM_OTHER << 8) | pdev->class;
768 +-}
769 +-DECLARE_PCI_FIXUP_CLASS_HEADER(APEX_PCI_VENDOR_ID, APEX_PCI_DEVICE_ID,
770 +- PCI_CLASS_NOT_DEFINED, 8, apex_pci_fixup_class);
771 +-
772 + static int apex_pci_probe(struct pci_dev *pci_dev,
773 + const struct pci_device_id *id)
774 + {
775 +diff --git a/drivers/target/target_core_fabric_lib.c b/drivers/target/target_core_fabric_lib.c
776 +index 10fae26b44ad..939c6212d2ac 100644
777 +--- a/drivers/target/target_core_fabric_lib.c
778 ++++ b/drivers/target/target_core_fabric_lib.c
779 +@@ -76,7 +76,7 @@ static int fc_get_pr_transport_id(
780 + * encoded TransportID.
781 + */
782 + ptr = &se_nacl->initiatorname[0];
783 +- for (i = 0; i < 24; ) {
784 ++ for (i = 0; i < 23; ) {
785 + if (!strncmp(&ptr[i], ":", 1)) {
786 + i++;
787 + continue;
788 +diff --git a/drivers/target/target_core_user.c b/drivers/target/target_core_user.c
789 +index 7ee0a75ce452..eff1e36ca03c 100644
790 +--- a/drivers/target/target_core_user.c
791 ++++ b/drivers/target/target_core_user.c
792 +@@ -2067,6 +2067,7 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
793 + mb->cmd_tail = 0;
794 + mb->cmd_head = 0;
795 + tcmu_flush_dcache_range(mb, sizeof(*mb));
796 ++ clear_bit(TCMU_DEV_BIT_BROKEN, &udev->flags);
797 +
798 + del_timer(&udev->cmd_timer);
799 +
800 +diff --git a/drivers/usb/dwc3/gadget.c b/drivers/usb/dwc3/gadget.c
801 +index 8222e674c777..1a6c973da487 100644
802 +--- a/drivers/usb/dwc3/gadget.c
803 ++++ b/drivers/usb/dwc3/gadget.c
804 +@@ -1577,7 +1577,6 @@ static int __dwc3_gadget_wakeup(struct dwc3 *dwc)
805 + u32 reg;
806 +
807 + u8 link_state;
808 +- u8 speed;
809 +
810 + /*
811 + * According to the Databook Remote wakeup request should
812 +@@ -1587,16 +1586,13 @@ static int __dwc3_gadget_wakeup(struct dwc3 *dwc)
813 + */
814 + reg = dwc3_readl(dwc->regs, DWC3_DSTS);
815 +
816 +- speed = reg & DWC3_DSTS_CONNECTSPD;
817 +- if ((speed == DWC3_DSTS_SUPERSPEED) ||
818 +- (speed == DWC3_DSTS_SUPERSPEED_PLUS))
819 +- return 0;
820 +-
821 + link_state = DWC3_DSTS_USBLNKST(reg);
822 +
823 + switch (link_state) {
824 ++ case DWC3_LINK_STATE_RESET:
825 + case DWC3_LINK_STATE_RX_DET: /* in HS, means Early Suspend */
826 + case DWC3_LINK_STATE_U3: /* in HS, means SUSPEND */
827 ++ case DWC3_LINK_STATE_RESUME:
828 + break;
829 + default:
830 + return -EINVAL;
831 +diff --git a/drivers/usb/gadget/udc/bdc/bdc_ep.c b/drivers/usb/gadget/udc/bdc/bdc_ep.c
832 +index a4d9b5e1e50e..d49c6dc1082d 100644
833 +--- a/drivers/usb/gadget/udc/bdc/bdc_ep.c
834 ++++ b/drivers/usb/gadget/udc/bdc/bdc_ep.c
835 +@@ -540,7 +540,7 @@ static void bdc_req_complete(struct bdc_ep *ep, struct bdc_req *req,
836 + {
837 + struct bdc *bdc = ep->bdc;
838 +
839 +- if (req == NULL || &req->queue == NULL || &req->usb_req == NULL)
840 ++ if (req == NULL)
841 + return;
842 +
843 + dev_dbg(bdc->dev, "%s ep:%s status:%d\n", __func__, ep->name, status);
844 +diff --git a/drivers/xen/xenbus/xenbus_client.c b/drivers/xen/xenbus/xenbus_client.c
845 +index a1c17000129b..e94a61eaeceb 100644
846 +--- a/drivers/xen/xenbus/xenbus_client.c
847 ++++ b/drivers/xen/xenbus/xenbus_client.c
848 +@@ -450,7 +450,14 @@ EXPORT_SYMBOL_GPL(xenbus_free_evtchn);
849 + int xenbus_map_ring_valloc(struct xenbus_device *dev, grant_ref_t *gnt_refs,
850 + unsigned int nr_grefs, void **vaddr)
851 + {
852 +- return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr);
853 ++ int err;
854 ++
855 ++ err = ring_ops->map(dev, gnt_refs, nr_grefs, vaddr);
856 ++ /* Some hypervisors are buggy and can return 1. */
857 ++ if (err > 0)
858 ++ err = GNTST_general_error;
859 ++
860 ++ return err;
861 + }
862 + EXPORT_SYMBOL_GPL(xenbus_map_ring_valloc);
863 +
864 +diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c
865 +index dafa7e4aaecb..8876eaad10f6 100644
866 +--- a/fs/ext4/ialloc.c
867 ++++ b/fs/ext4/ialloc.c
868 +@@ -665,7 +665,7 @@ static int find_group_other(struct super_block *sb, struct inode *parent,
869 + * block has been written back to disk. (Yes, these values are
870 + * somewhat arbitrary...)
871 + */
872 +-#define RECENTCY_MIN 5
873 ++#define RECENTCY_MIN 60
874 + #define RECENTCY_DIRTY 300
875 +
876 + static int recently_deleted(struct super_block *sb, ext4_group_t group, int ino)
877 +diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c
878 +index 000fa0e39278..3b1a7597af15 100644
879 +--- a/fs/ext4/inode.c
880 ++++ b/fs/ext4/inode.c
881 +@@ -2128,7 +2128,7 @@ static int ext4_writepage(struct page *page,
882 + bool keep_towrite = false;
883 +
884 + if (unlikely(ext4_forced_shutdown(EXT4_SB(inode->i_sb)))) {
885 +- ext4_invalidatepage(page, 0, PAGE_SIZE);
886 ++ inode->i_mapping->a_ops->invalidatepage(page, 0, PAGE_SIZE);
887 + unlock_page(page);
888 + return -EIO;
889 + }
890 +diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c
891 +index 71121fcf9e8c..8dd54a8a0361 100644
892 +--- a/fs/ext4/mballoc.c
893 ++++ b/fs/ext4/mballoc.c
894 +@@ -1936,7 +1936,8 @@ void ext4_mb_complex_scan_group(struct ext4_allocation_context *ac,
895 + int free;
896 +
897 + free = e4b->bd_info->bb_free;
898 +- BUG_ON(free <= 0);
899 ++ if (WARN_ON(free <= 0))
900 ++ return;
901 +
902 + i = e4b->bd_info->bb_first_free;
903 +
904 +@@ -1959,7 +1960,8 @@ void ext4_mb_complex_scan_group(struct ext4_allocation_context *ac,
905 + }
906 +
907 + mb_find_extent(e4b, i, ac->ac_g_ex.fe_len, &ex);
908 +- BUG_ON(ex.fe_len <= 0);
909 ++ if (WARN_ON(ex.fe_len <= 0))
910 ++ break;
911 + if (free < ex.fe_len) {
912 + ext4_grp_locked_error(sb, e4b->bd_group, 0, 0,
913 + "%d free clusters as per "
914 +diff --git a/fs/ext4/super.c b/fs/ext4/super.c
915 +index bf949fcc970a..93c14ecac831 100644
916 +--- a/fs/ext4/super.c
917 ++++ b/fs/ext4/super.c
918 +@@ -3524,7 +3524,8 @@ int ext4_calculate_overhead(struct super_block *sb)
919 + */
920 + if (sbi->s_journal && !sbi->journal_bdev)
921 + overhead += EXT4_NUM_B2C(sbi, sbi->s_journal->j_maxlen);
922 +- else if (ext4_has_feature_journal(sb) && !sbi->s_journal) {
923 ++ else if (ext4_has_feature_journal(sb) && !sbi->s_journal && j_inum) {
924 ++ /* j_inum for internal journal is non-zero */
925 + j_inode = ext4_get_journal_inode(sb, j_inum);
926 + if (j_inode) {
927 + j_blocks = j_inode->i_size >> sb->s_blocksize_bits;
928 +diff --git a/fs/nfsd/nfs4state.c b/fs/nfsd/nfs4state.c
929 +index ed73e86194fa..c24306af9758 100644
930 +--- a/fs/nfsd/nfs4state.c
931 ++++ b/fs/nfsd/nfs4state.c
932 +@@ -252,6 +252,8 @@ find_or_allocate_block(struct nfs4_lockowner *lo, struct knfsd_fh *fh,
933 + if (!nbl) {
934 + nbl= kmalloc(sizeof(*nbl), GFP_KERNEL);
935 + if (nbl) {
936 ++ INIT_LIST_HEAD(&nbl->nbl_list);
937 ++ INIT_LIST_HEAD(&nbl->nbl_lru);
938 + fh_copy_shallow(&nbl->nbl_fh, fh);
939 + locks_init_lock(&nbl->nbl_lock);
940 + nfsd4_init_cb(&nbl->nbl_cb, lo->lo_owner.so_client,
941 +diff --git a/fs/pnode.c b/fs/pnode.c
942 +index 53d411a371ce..7910ae91f17e 100644
943 +--- a/fs/pnode.c
944 ++++ b/fs/pnode.c
945 +@@ -266,14 +266,13 @@ static int propagate_one(struct mount *m)
946 + if (IS_ERR(child))
947 + return PTR_ERR(child);
948 + child->mnt.mnt_flags &= ~MNT_LOCKED;
949 ++ read_seqlock_excl(&mount_lock);
950 + mnt_set_mountpoint(m, mp, child);
951 ++ if (m->mnt_master != dest_master)
952 ++ SET_MNT_MARK(m->mnt_master);
953 ++ read_sequnlock_excl(&mount_lock);
954 + last_dest = m;
955 + last_source = child;
956 +- if (m->mnt_master != dest_master) {
957 +- read_seqlock_excl(&mount_lock);
958 +- SET_MNT_MARK(m->mnt_master);
959 +- read_sequnlock_excl(&mount_lock);
960 +- }
961 + hlist_add_head(&child->mnt_hash, list);
962 + return count_mounts(m->mnt_ns, child);
963 + }
964 +diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c
965 +index 245483cc282b..901f27ac94ab 100644
966 +--- a/fs/xfs/xfs_icache.c
967 ++++ b/fs/xfs/xfs_icache.c
968 +@@ -902,7 +902,12 @@ xfs_eofblocks_worker(
969 + {
970 + struct xfs_mount *mp = container_of(to_delayed_work(work),
971 + struct xfs_mount, m_eofblocks_work);
972 ++
973 ++ if (!sb_start_write_trylock(mp->m_super))
974 ++ return;
975 + xfs_icache_free_eofblocks(mp, NULL);
976 ++ sb_end_write(mp->m_super);
977 ++
978 + xfs_queue_eofblocks(mp);
979 + }
980 +
981 +@@ -929,7 +934,12 @@ xfs_cowblocks_worker(
982 + {
983 + struct xfs_mount *mp = container_of(to_delayed_work(work),
984 + struct xfs_mount, m_cowblocks_work);
985 ++
986 ++ if (!sb_start_write_trylock(mp->m_super))
987 ++ return;
988 + xfs_icache_free_cowblocks(mp, NULL);
989 ++ sb_end_write(mp->m_super);
990 ++
991 + xfs_queue_cowblocks(mp);
992 + }
993 +
994 +diff --git a/fs/xfs/xfs_ioctl.c b/fs/xfs/xfs_ioctl.c
995 +index bad90479ade2..6ffb53edf1b4 100644
996 +--- a/fs/xfs/xfs_ioctl.c
997 ++++ b/fs/xfs/xfs_ioctl.c
998 +@@ -2182,7 +2182,10 @@ xfs_file_ioctl(
999 + if (error)
1000 + return error;
1001 +
1002 +- return xfs_icache_free_eofblocks(mp, &keofb);
1003 ++ sb_start_write(mp->m_super);
1004 ++ error = xfs_icache_free_eofblocks(mp, &keofb);
1005 ++ sb_end_write(mp->m_super);
1006 ++ return error;
1007 + }
1008 +
1009 + default:
1010 +diff --git a/fs/xfs/xfs_reflink.c b/fs/xfs/xfs_reflink.c
1011 +index f3c393f309e1..6622652a85a8 100644
1012 +--- a/fs/xfs/xfs_reflink.c
1013 ++++ b/fs/xfs/xfs_reflink.c
1014 +@@ -1058,6 +1058,7 @@ xfs_reflink_remap_extent(
1015 + uirec.br_startblock = irec->br_startblock + rlen;
1016 + uirec.br_startoff = irec->br_startoff + rlen;
1017 + uirec.br_blockcount = unmap_len - rlen;
1018 ++ uirec.br_state = irec->br_state;
1019 + unmap_len = rlen;
1020 +
1021 + /* If this isn't a real mapping, we're done. */
1022 +diff --git a/fs/xfs/xfs_trans_ail.c b/fs/xfs/xfs_trans_ail.c
1023 +index d3a4e89bf4a0..66f167aefd94 100644
1024 +--- a/fs/xfs/xfs_trans_ail.c
1025 ++++ b/fs/xfs/xfs_trans_ail.c
1026 +@@ -520,8 +520,9 @@ xfsaild(
1027 + {
1028 + struct xfs_ail *ailp = data;
1029 + long tout = 0; /* milliseconds */
1030 ++ unsigned int noreclaim_flag;
1031 +
1032 +- current->flags |= PF_MEMALLOC;
1033 ++ noreclaim_flag = memalloc_noreclaim_save();
1034 + set_freezable();
1035 +
1036 + while (1) {
1037 +@@ -592,6 +593,7 @@ xfsaild(
1038 + tout = xfsaild_push(ailp);
1039 + }
1040 +
1041 ++ memalloc_noreclaim_restore(noreclaim_flag);
1042 + return 0;
1043 + }
1044 +
1045 +diff --git a/include/linux/qed/qed_chain.h b/include/linux/qed/qed_chain.h
1046 +index 2dd0a9ed5b36..733fad7dfbed 100644
1047 +--- a/include/linux/qed/qed_chain.h
1048 ++++ b/include/linux/qed/qed_chain.h
1049 +@@ -97,6 +97,11 @@ struct qed_chain_u32 {
1050 + u32 cons_idx;
1051 + };
1052 +
1053 ++struct addr_tbl_entry {
1054 ++ void *virt_addr;
1055 ++ dma_addr_t dma_map;
1056 ++};
1057 ++
1058 + struct qed_chain {
1059 + /* fastpath portion of the chain - required for commands such
1060 + * as produce / consume.
1061 +@@ -107,10 +112,11 @@ struct qed_chain {
1062 +
1063 + /* Fastpath portions of the PBL [if exists] */
1064 + struct {
1065 +- /* Table for keeping the virtual addresses of the chain pages,
1066 +- * respectively to the physical addresses in the pbl table.
1067 ++ /* Table for keeping the virtual and physical addresses of the
1068 ++ * chain pages, respectively to the physical addresses
1069 ++ * in the pbl table.
1070 + */
1071 +- void **pp_virt_addr_tbl;
1072 ++ struct addr_tbl_entry *pp_addr_tbl;
1073 +
1074 + union {
1075 + struct qed_chain_pbl_u16 u16;
1076 +@@ -287,7 +293,7 @@ qed_chain_advance_page(struct qed_chain *p_chain,
1077 + *(u32 *)page_to_inc = 0;
1078 + page_index = *(u32 *)page_to_inc;
1079 + }
1080 +- *p_next_elem = p_chain->pbl.pp_virt_addr_tbl[page_index];
1081 ++ *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr;
1082 + }
1083 + }
1084 +
1085 +@@ -537,7 +543,7 @@ static inline void qed_chain_init_params(struct qed_chain *p_chain,
1086 +
1087 + p_chain->pbl_sp.p_phys_table = 0;
1088 + p_chain->pbl_sp.p_virt_table = NULL;
1089 +- p_chain->pbl.pp_virt_addr_tbl = NULL;
1090 ++ p_chain->pbl.pp_addr_tbl = NULL;
1091 + }
1092 +
1093 + /**
1094 +@@ -575,11 +581,11 @@ static inline void qed_chain_init_mem(struct qed_chain *p_chain,
1095 + static inline void qed_chain_init_pbl_mem(struct qed_chain *p_chain,
1096 + void *p_virt_pbl,
1097 + dma_addr_t p_phys_pbl,
1098 +- void **pp_virt_addr_tbl)
1099 ++ struct addr_tbl_entry *pp_addr_tbl)
1100 + {
1101 + p_chain->pbl_sp.p_phys_table = p_phys_pbl;
1102 + p_chain->pbl_sp.p_virt_table = p_virt_pbl;
1103 +- p_chain->pbl.pp_virt_addr_tbl = pp_virt_addr_tbl;
1104 ++ p_chain->pbl.pp_addr_tbl = pp_addr_tbl;
1105 + }
1106 +
1107 + /**
1108 +@@ -644,7 +650,7 @@ static inline void *qed_chain_get_last_elem(struct qed_chain *p_chain)
1109 + break;
1110 + case QED_CHAIN_MODE_PBL:
1111 + last_page_idx = p_chain->page_cnt - 1;
1112 +- p_virt_addr = p_chain->pbl.pp_virt_addr_tbl[last_page_idx];
1113 ++ p_virt_addr = p_chain->pbl.pp_addr_tbl[last_page_idx].virt_addr;
1114 + break;
1115 + }
1116 + /* p_virt_addr points at this stage to the last page of the chain */
1117 +@@ -716,7 +722,7 @@ static inline void qed_chain_pbl_zero_mem(struct qed_chain *p_chain)
1118 + page_cnt = qed_chain_get_page_cnt(p_chain);
1119 +
1120 + for (i = 0; i < page_cnt; i++)
1121 +- memset(p_chain->pbl.pp_virt_addr_tbl[i], 0,
1122 ++ memset(p_chain->pbl.pp_addr_tbl[i].virt_addr, 0,
1123 + QED_CHAIN_PAGE_SIZE);
1124 + }
1125 +
1126 +diff --git a/include/linux/sunrpc/svc_rdma.h b/include/linux/sunrpc/svc_rdma.h
1127 +index fd78f78df5c6..3e3214cca447 100644
1128 +--- a/include/linux/sunrpc/svc_rdma.h
1129 ++++ b/include/linux/sunrpc/svc_rdma.h
1130 +@@ -159,6 +159,7 @@ extern bool svc_rdma_post_recvs(struct svcxprt_rdma *rdma);
1131 + extern void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma,
1132 + struct svc_rdma_recv_ctxt *ctxt);
1133 + extern void svc_rdma_flush_recv_queues(struct svcxprt_rdma *rdma);
1134 ++extern void svc_rdma_release_rqst(struct svc_rqst *rqstp);
1135 + extern int svc_rdma_recvfrom(struct svc_rqst *);
1136 +
1137 + /* svc_rdma_rw.c */
1138 +diff --git a/include/trace/events/rpcrdma.h b/include/trace/events/rpcrdma.h
1139 +index 4c91cadd1871..4a0b18921e06 100644
1140 +--- a/include/trace/events/rpcrdma.h
1141 ++++ b/include/trace/events/rpcrdma.h
1142 +@@ -1322,17 +1322,15 @@ DECLARE_EVENT_CLASS(svcrdma_sendcomp_event,
1143 +
1144 + TRACE_EVENT(svcrdma_post_send,
1145 + TP_PROTO(
1146 +- const struct ib_send_wr *wr,
1147 +- int status
1148 ++ const struct ib_send_wr *wr
1149 + ),
1150 +
1151 +- TP_ARGS(wr, status),
1152 ++ TP_ARGS(wr),
1153 +
1154 + TP_STRUCT__entry(
1155 + __field(const void *, cqe)
1156 + __field(unsigned int, num_sge)
1157 + __field(u32, inv_rkey)
1158 +- __field(int, status)
1159 + ),
1160 +
1161 + TP_fast_assign(
1162 +@@ -1340,12 +1338,11 @@ TRACE_EVENT(svcrdma_post_send,
1163 + __entry->num_sge = wr->num_sge;
1164 + __entry->inv_rkey = (wr->opcode == IB_WR_SEND_WITH_INV) ?
1165 + wr->ex.invalidate_rkey : 0;
1166 +- __entry->status = status;
1167 + ),
1168 +
1169 +- TP_printk("cqe=%p num_sge=%u inv_rkey=0x%08x status=%d",
1170 ++ TP_printk("cqe=%p num_sge=%u inv_rkey=0x%08x",
1171 + __entry->cqe, __entry->num_sge,
1172 +- __entry->inv_rkey, __entry->status
1173 ++ __entry->inv_rkey
1174 + )
1175 + );
1176 +
1177 +@@ -1410,26 +1407,23 @@ TRACE_EVENT(svcrdma_wc_receive,
1178 + TRACE_EVENT(svcrdma_post_rw,
1179 + TP_PROTO(
1180 + const void *cqe,
1181 +- int sqecount,
1182 +- int status
1183 ++ int sqecount
1184 + ),
1185 +
1186 +- TP_ARGS(cqe, sqecount, status),
1187 ++ TP_ARGS(cqe, sqecount),
1188 +
1189 + TP_STRUCT__entry(
1190 + __field(const void *, cqe)
1191 + __field(int, sqecount)
1192 +- __field(int, status)
1193 + ),
1194 +
1195 + TP_fast_assign(
1196 + __entry->cqe = cqe;
1197 + __entry->sqecount = sqecount;
1198 +- __entry->status = status;
1199 + ),
1200 +
1201 +- TP_printk("cqe=%p sqecount=%d status=%d",
1202 +- __entry->cqe, __entry->sqecount, __entry->status
1203 ++ TP_printk("cqe=%p sqecount=%d",
1204 ++ __entry->cqe, __entry->sqecount
1205 + )
1206 + );
1207 +
1208 +@@ -1525,6 +1519,34 @@ DECLARE_EVENT_CLASS(svcrdma_sendqueue_event,
1209 + DEFINE_SQ_EVENT(full);
1210 + DEFINE_SQ_EVENT(retry);
1211 +
1212 ++TRACE_EVENT(svcrdma_sq_post_err,
1213 ++ TP_PROTO(
1214 ++ const struct svcxprt_rdma *rdma,
1215 ++ int status
1216 ++ ),
1217 ++
1218 ++ TP_ARGS(rdma, status),
1219 ++
1220 ++ TP_STRUCT__entry(
1221 ++ __field(int, avail)
1222 ++ __field(int, depth)
1223 ++ __field(int, status)
1224 ++ __string(addr, rdma->sc_xprt.xpt_remotebuf)
1225 ++ ),
1226 ++
1227 ++ TP_fast_assign(
1228 ++ __entry->avail = atomic_read(&rdma->sc_sq_avail);
1229 ++ __entry->depth = rdma->sc_sq_depth;
1230 ++ __entry->status = status;
1231 ++ __assign_str(addr, rdma->sc_xprt.xpt_remotebuf);
1232 ++ ),
1233 ++
1234 ++ TP_printk("addr=%s sc_sq_avail=%d/%d status=%d",
1235 ++ __get_str(addr), __entry->avail, __entry->depth,
1236 ++ __entry->status
1237 ++ )
1238 ++);
1239 ++
1240 + #endif /* _TRACE_RPCRDMA_H */
1241 +
1242 + #include <trace/define_trace.h>
1243 +diff --git a/include/uapi/linux/swab.h b/include/uapi/linux/swab.h
1244 +index fa7f97da5b76..7272f85d6d6a 100644
1245 +--- a/include/uapi/linux/swab.h
1246 ++++ b/include/uapi/linux/swab.h
1247 +@@ -135,9 +135,9 @@ static inline __attribute_const__ __u32 __fswahb32(__u32 val)
1248 +
1249 + static __always_inline unsigned long __swab(const unsigned long y)
1250 + {
1251 +-#if BITS_PER_LONG == 64
1252 ++#if __BITS_PER_LONG == 64
1253 + return __swab64(y);
1254 +-#else /* BITS_PER_LONG == 32 */
1255 ++#else /* __BITS_PER_LONG == 32 */
1256 + return __swab32(y);
1257 + #endif
1258 + }
1259 +diff --git a/kernel/bpf/cpumap.c b/kernel/bpf/cpumap.c
1260 +index 3c18260403dd..61fbcae82f0a 100644
1261 +--- a/kernel/bpf/cpumap.c
1262 ++++ b/kernel/bpf/cpumap.c
1263 +@@ -455,7 +455,7 @@ static int cpu_map_update_elem(struct bpf_map *map, void *key, void *value,
1264 + return -EOVERFLOW;
1265 +
1266 + /* Make sure CPU is a valid possible cpu */
1267 +- if (!cpu_possible(key_cpu))
1268 ++ if (key_cpu >= nr_cpumask_bits || !cpu_possible(key_cpu))
1269 + return -ENODEV;
1270 +
1271 + if (qsize == 0) {
1272 +diff --git a/kernel/events/core.c b/kernel/events/core.c
1273 +index 00fb2fe92c4d..21e3c65abc76 100644
1274 +--- a/kernel/events/core.c
1275 ++++ b/kernel/events/core.c
1276 +@@ -6923,10 +6923,17 @@ static void perf_event_task_output(struct perf_event *event,
1277 + goto out;
1278 +
1279 + task_event->event_id.pid = perf_event_pid(event, task);
1280 +- task_event->event_id.ppid = perf_event_pid(event, current);
1281 +-
1282 + task_event->event_id.tid = perf_event_tid(event, task);
1283 +- task_event->event_id.ptid = perf_event_tid(event, current);
1284 ++
1285 ++ if (task_event->event_id.header.type == PERF_RECORD_EXIT) {
1286 ++ task_event->event_id.ppid = perf_event_pid(event,
1287 ++ task->real_parent);
1288 ++ task_event->event_id.ptid = perf_event_pid(event,
1289 ++ task->real_parent);
1290 ++ } else { /* PERF_RECORD_FORK */
1291 ++ task_event->event_id.ppid = perf_event_pid(event, current);
1292 ++ task_event->event_id.ptid = perf_event_tid(event, current);
1293 ++ }
1294 +
1295 + task_event->event_id.time = perf_event_clock(event);
1296 +
1297 +diff --git a/mm/shmem.c b/mm/shmem.c
1298 +index a6fc82a1ab6b..0a2e1e7801d6 100644
1299 +--- a/mm/shmem.c
1300 ++++ b/mm/shmem.c
1301 +@@ -2350,11 +2350,11 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
1302 +
1303 + lru_cache_add_anon(page);
1304 +
1305 +- spin_lock(&info->lock);
1306 ++ spin_lock_irq(&info->lock);
1307 + info->alloced++;
1308 + inode->i_blocks += BLOCKS_PER_PAGE;
1309 + shmem_recalc_inode(inode);
1310 +- spin_unlock(&info->lock);
1311 ++ spin_unlock_irq(&info->lock);
1312 +
1313 + inc_mm_counter(dst_mm, mm_counter_file(page));
1314 + page_add_file_rmap(page, false);
1315 +diff --git a/net/rxrpc/local_object.c b/net/rxrpc/local_object.c
1316 +index 4c0087a48e87..fe190a691872 100644
1317 +--- a/net/rxrpc/local_object.c
1318 ++++ b/net/rxrpc/local_object.c
1319 +@@ -169,15 +169,6 @@ static int rxrpc_open_socket(struct rxrpc_local *local, struct net *net)
1320 + goto error;
1321 + }
1322 +
1323 +- /* we want to set the don't fragment bit */
1324 +- opt = IPV6_PMTUDISC_DO;
1325 +- ret = kernel_setsockopt(local->socket, SOL_IPV6, IPV6_MTU_DISCOVER,
1326 +- (char *) &opt, sizeof(opt));
1327 +- if (ret < 0) {
1328 +- _debug("setsockopt failed");
1329 +- goto error;
1330 +- }
1331 +-
1332 + /* Fall through and set IPv4 options too otherwise we don't get
1333 + * errors from IPv4 packets sent through the IPv6 socket.
1334 + */
1335 +diff --git a/net/rxrpc/output.c b/net/rxrpc/output.c
1336 +index b0aa08e3796d..da8a555ec3f6 100644
1337 +--- a/net/rxrpc/output.c
1338 ++++ b/net/rxrpc/output.c
1339 +@@ -480,41 +480,21 @@ send_fragmentable:
1340 + skb->tstamp = ktime_get_real();
1341 +
1342 + switch (conn->params.local->srx.transport.family) {
1343 ++ case AF_INET6:
1344 + case AF_INET:
1345 + opt = IP_PMTUDISC_DONT;
1346 +- ret = kernel_setsockopt(conn->params.local->socket,
1347 +- SOL_IP, IP_MTU_DISCOVER,
1348 +- (char *)&opt, sizeof(opt));
1349 +- if (ret == 0) {
1350 +- ret = kernel_sendmsg(conn->params.local->socket, &msg,
1351 +- iov, 2, len);
1352 +- conn->params.peer->last_tx_at = ktime_get_seconds();
1353 +-
1354 +- opt = IP_PMTUDISC_DO;
1355 +- kernel_setsockopt(conn->params.local->socket, SOL_IP,
1356 +- IP_MTU_DISCOVER,
1357 +- (char *)&opt, sizeof(opt));
1358 +- }
1359 +- break;
1360 +-
1361 +-#ifdef CONFIG_AF_RXRPC_IPV6
1362 +- case AF_INET6:
1363 +- opt = IPV6_PMTUDISC_DONT;
1364 +- ret = kernel_setsockopt(conn->params.local->socket,
1365 +- SOL_IPV6, IPV6_MTU_DISCOVER,
1366 +- (char *)&opt, sizeof(opt));
1367 +- if (ret == 0) {
1368 +- ret = kernel_sendmsg(conn->params.local->socket, &msg,
1369 +- iov, 2, len);
1370 +- conn->params.peer->last_tx_at = ktime_get_seconds();
1371 +-
1372 +- opt = IPV6_PMTUDISC_DO;
1373 +- kernel_setsockopt(conn->params.local->socket,
1374 +- SOL_IPV6, IPV6_MTU_DISCOVER,
1375 +- (char *)&opt, sizeof(opt));
1376 +- }
1377 ++ kernel_setsockopt(conn->params.local->socket,
1378 ++ SOL_IP, IP_MTU_DISCOVER,
1379 ++ (char *)&opt, sizeof(opt));
1380 ++ ret = kernel_sendmsg(conn->params.local->socket, &msg,
1381 ++ iov, 2, len);
1382 ++ conn->params.peer->last_tx_at = ktime_get_seconds();
1383 ++
1384 ++ opt = IP_PMTUDISC_DO;
1385 ++ kernel_setsockopt(conn->params.local->socket,
1386 ++ SOL_IP, IP_MTU_DISCOVER,
1387 ++ (char *)&opt, sizeof(opt));
1388 + break;
1389 +-#endif
1390 +
1391 + default:
1392 + BUG();
1393 +diff --git a/net/sunrpc/svc_xprt.c b/net/sunrpc/svc_xprt.c
1394 +index 6cf0fd37cbf0..c8ee8e801edb 100644
1395 +--- a/net/sunrpc/svc_xprt.c
1396 ++++ b/net/sunrpc/svc_xprt.c
1397 +@@ -878,9 +878,6 @@ int svc_send(struct svc_rqst *rqstp)
1398 + if (!xprt)
1399 + goto out;
1400 +
1401 +- /* release the receive skb before sending the reply */
1402 +- xprt->xpt_ops->xpo_release_rqst(rqstp);
1403 +-
1404 + /* calculate over-all length */
1405 + xb = &rqstp->rq_res;
1406 + xb->len = xb->head[0].iov_len +
1407 +diff --git a/net/sunrpc/svcsock.c b/net/sunrpc/svcsock.c
1408 +index 8566531c2f10..d0b5a1c47a32 100644
1409 +--- a/net/sunrpc/svcsock.c
1410 ++++ b/net/sunrpc/svcsock.c
1411 +@@ -636,6 +636,8 @@ svc_udp_sendto(struct svc_rqst *rqstp)
1412 + {
1413 + int error;
1414 +
1415 ++ svc_release_udp_skb(rqstp);
1416 ++
1417 + error = svc_sendto(rqstp, &rqstp->rq_res);
1418 + if (error == -ECONNREFUSED)
1419 + /* ICMP error on earlier request. */
1420 +@@ -1173,6 +1175,8 @@ static int svc_tcp_sendto(struct svc_rqst *rqstp)
1421 + int sent;
1422 + __be32 reclen;
1423 +
1424 ++ svc_release_skb(rqstp);
1425 ++
1426 + /* Set up the first element of the reply kvec.
1427 + * Any other kvecs that may be in use have been taken
1428 + * care of by the server implementation itself.
1429 +diff --git a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
1430 +index b24d5b8f2fee..16c8174658fd 100644
1431 +--- a/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
1432 ++++ b/net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
1433 +@@ -226,6 +226,26 @@ void svc_rdma_recv_ctxt_put(struct svcxprt_rdma *rdma,
1434 + svc_rdma_recv_ctxt_destroy(rdma, ctxt);
1435 + }
1436 +
1437 ++/**
1438 ++ * svc_rdma_release_rqst - Release transport-specific per-rqst resources
1439 ++ * @rqstp: svc_rqst being released
1440 ++ *
1441 ++ * Ensure that the recv_ctxt is released whether or not a Reply
1442 ++ * was sent. For example, the client could close the connection,
1443 ++ * or svc_process could drop an RPC, before the Reply is sent.
1444 ++ */
1445 ++void svc_rdma_release_rqst(struct svc_rqst *rqstp)
1446 ++{
1447 ++ struct svc_rdma_recv_ctxt *ctxt = rqstp->rq_xprt_ctxt;
1448 ++ struct svc_xprt *xprt = rqstp->rq_xprt;
1449 ++ struct svcxprt_rdma *rdma =
1450 ++ container_of(xprt, struct svcxprt_rdma, sc_xprt);
1451 ++
1452 ++ rqstp->rq_xprt_ctxt = NULL;
1453 ++ if (ctxt)
1454 ++ svc_rdma_recv_ctxt_put(rdma, ctxt);
1455 ++}
1456 ++
1457 + static int __svc_rdma_post_recv(struct svcxprt_rdma *rdma,
1458 + struct svc_rdma_recv_ctxt *ctxt)
1459 + {
1460 +@@ -704,6 +724,8 @@ int svc_rdma_recvfrom(struct svc_rqst *rqstp)
1461 + __be32 *p;
1462 + int ret;
1463 +
1464 ++ rqstp->rq_xprt_ctxt = NULL;
1465 ++
1466 + spin_lock(&rdma_xprt->sc_rq_dto_lock);
1467 + ctxt = svc_rdma_next_recv_ctxt(&rdma_xprt->sc_read_complete_q);
1468 + if (ctxt) {
1469 +diff --git a/net/sunrpc/xprtrdma/svc_rdma_rw.c b/net/sunrpc/xprtrdma/svc_rdma_rw.c
1470 +index dc1951759a8e..4fc0ce127089 100644
1471 +--- a/net/sunrpc/xprtrdma/svc_rdma_rw.c
1472 ++++ b/net/sunrpc/xprtrdma/svc_rdma_rw.c
1473 +@@ -331,8 +331,6 @@ static int svc_rdma_post_chunk_ctxt(struct svc_rdma_chunk_ctxt *cc)
1474 + if (atomic_sub_return(cc->cc_sqecount,
1475 + &rdma->sc_sq_avail) > 0) {
1476 + ret = ib_post_send(rdma->sc_qp, first_wr, &bad_wr);
1477 +- trace_svcrdma_post_rw(&cc->cc_cqe,
1478 +- cc->cc_sqecount, ret);
1479 + if (ret)
1480 + break;
1481 + return 0;
1482 +@@ -345,6 +343,7 @@ static int svc_rdma_post_chunk_ctxt(struct svc_rdma_chunk_ctxt *cc)
1483 + trace_svcrdma_sq_retry(rdma);
1484 + } while (1);
1485 +
1486 ++ trace_svcrdma_sq_post_err(rdma, ret);
1487 + set_bit(XPT_CLOSE, &xprt->xpt_flags);
1488 +
1489 + /* If even one was posted, there will be a completion. */
1490 +diff --git a/net/sunrpc/xprtrdma/svc_rdma_sendto.c b/net/sunrpc/xprtrdma/svc_rdma_sendto.c
1491 +index e8ad7ddf347a..aa4d19a780d7 100644
1492 +--- a/net/sunrpc/xprtrdma/svc_rdma_sendto.c
1493 ++++ b/net/sunrpc/xprtrdma/svc_rdma_sendto.c
1494 +@@ -310,15 +310,17 @@ int svc_rdma_send(struct svcxprt_rdma *rdma, struct ib_send_wr *wr)
1495 + }
1496 +
1497 + svc_xprt_get(&rdma->sc_xprt);
1498 ++ trace_svcrdma_post_send(wr);
1499 + ret = ib_post_send(rdma->sc_qp, wr, NULL);
1500 +- trace_svcrdma_post_send(wr, ret);
1501 +- if (ret) {
1502 +- set_bit(XPT_CLOSE, &rdma->sc_xprt.xpt_flags);
1503 +- svc_xprt_put(&rdma->sc_xprt);
1504 +- wake_up(&rdma->sc_send_wait);
1505 +- }
1506 +- break;
1507 ++ if (ret)
1508 ++ break;
1509 ++ return 0;
1510 + }
1511 ++
1512 ++ trace_svcrdma_sq_post_err(rdma, ret);
1513 ++ set_bit(XPT_CLOSE, &rdma->sc_xprt.xpt_flags);
1514 ++ svc_xprt_put(&rdma->sc_xprt);
1515 ++ wake_up(&rdma->sc_send_wait);
1516 + return ret;
1517 + }
1518 +
1519 +@@ -906,12 +908,7 @@ int svc_rdma_sendto(struct svc_rqst *rqstp)
1520 + wr_lst, rp_ch);
1521 + if (ret < 0)
1522 + goto err1;
1523 +- ret = 0;
1524 +-
1525 +-out:
1526 +- rqstp->rq_xprt_ctxt = NULL;
1527 +- svc_rdma_recv_ctxt_put(rdma, rctxt);
1528 +- return ret;
1529 ++ return 0;
1530 +
1531 + err2:
1532 + if (ret != -E2BIG && ret != -EINVAL)
1533 +@@ -920,14 +917,12 @@ out:
1534 + ret = svc_rdma_send_error_msg(rdma, sctxt, rqstp);
1535 + if (ret < 0)
1536 + goto err1;
1537 +- ret = 0;
1538 +- goto out;
1539 ++ return 0;
1540 +
1541 + err1:
1542 + svc_rdma_send_ctxt_put(rdma, sctxt);
1543 + err0:
1544 + trace_svcrdma_send_failed(rqstp, ret);
1545 + set_bit(XPT_CLOSE, &xprt->xpt_flags);
1546 +- ret = -ENOTCONN;
1547 +- goto out;
1548 ++ return -ENOTCONN;
1549 + }
1550 +diff --git a/net/sunrpc/xprtrdma/svc_rdma_transport.c b/net/sunrpc/xprtrdma/svc_rdma_transport.c
1551 +index 7308992b7a18..f1824303e076 100644
1552 +--- a/net/sunrpc/xprtrdma/svc_rdma_transport.c
1553 ++++ b/net/sunrpc/xprtrdma/svc_rdma_transport.c
1554 +@@ -71,7 +71,6 @@ static struct svc_xprt *svc_rdma_create(struct svc_serv *serv,
1555 + struct sockaddr *sa, int salen,
1556 + int flags);
1557 + static struct svc_xprt *svc_rdma_accept(struct svc_xprt *xprt);
1558 +-static void svc_rdma_release_rqst(struct svc_rqst *);
1559 + static void svc_rdma_detach(struct svc_xprt *xprt);
1560 + static void svc_rdma_free(struct svc_xprt *xprt);
1561 + static int svc_rdma_has_wspace(struct svc_xprt *xprt);
1562 +@@ -616,10 +615,6 @@ static struct svc_xprt *svc_rdma_accept(struct svc_xprt *xprt)
1563 + return NULL;
1564 + }
1565 +
1566 +-static void svc_rdma_release_rqst(struct svc_rqst *rqstp)
1567 +-{
1568 +-}
1569 +-
1570 + /*
1571 + * When connected, an svc_xprt has at least two references:
1572 + *
1573 +diff --git a/sound/pci/hda/hda_intel.c b/sound/pci/hda/hda_intel.c
1574 +index 0502042c1616..0d7981eda2c4 100644
1575 +--- a/sound/pci/hda/hda_intel.c
1576 ++++ b/sound/pci/hda/hda_intel.c
1577 +@@ -2054,7 +2054,7 @@ static int azx_first_init(struct azx *chip)
1578 + /* codec detection */
1579 + if (!azx_bus(chip)->codec_mask) {
1580 + dev_err(card->dev, "no codecs found!\n");
1581 +- return -ENODEV;
1582 ++ /* keep running the rest for the runtime PM */
1583 + }
1584 +
1585 + if (azx_acquire_irq(chip, 0) < 0)
1586 +@@ -2440,9 +2440,11 @@ static int azx_probe_continue(struct azx *chip)
1587 + #endif
1588 +
1589 + /* create codec instances */
1590 +- err = azx_probe_codecs(chip, azx_max_codecs[chip->driver_type]);
1591 +- if (err < 0)
1592 +- goto out_free;
1593 ++ if (bus->codec_mask) {
1594 ++ err = azx_probe_codecs(chip, azx_max_codecs[chip->driver_type]);
1595 ++ if (err < 0)
1596 ++ goto out_free;
1597 ++ }
1598 +
1599 + #ifdef CONFIG_SND_HDA_PATCH_LOADER
1600 + if (chip->fw) {
1601 +@@ -2456,7 +2458,7 @@ static int azx_probe_continue(struct azx *chip)
1602 + #endif
1603 + }
1604 + #endif
1605 +- if ((probe_only[dev] & 1) == 0) {
1606 ++ if (bus->codec_mask && !(probe_only[dev] & 1)) {
1607 + err = azx_codec_configure(chip);
1608 + if (err < 0)
1609 + goto out_free;
1610 +@@ -2473,8 +2475,11 @@ static int azx_probe_continue(struct azx *chip)
1611 +
1612 + set_default_power_save(chip);
1613 +
1614 +- if (azx_has_pm_runtime(chip))
1615 ++ if (azx_has_pm_runtime(chip)) {
1616 ++ pm_runtime_use_autosuspend(&pci->dev);
1617 ++ pm_runtime_allow(&pci->dev);
1618 + pm_runtime_put_autosuspend(&pci->dev);
1619 ++ }
1620 +
1621 + out_free:
1622 + if ((chip->driver_caps & AZX_DCAPS_I915_POWERWELL)
1623 +diff --git a/sound/soc/codecs/tas571x.c b/sound/soc/codecs/tas571x.c
1624 +index ca2dfe12344e..354f17a2fa5c 100644
1625 +--- a/sound/soc/codecs/tas571x.c
1626 ++++ b/sound/soc/codecs/tas571x.c
1627 +@@ -824,8 +824,10 @@ static int tas571x_i2c_probe(struct i2c_client *client,
1628 +
1629 + priv->regmap = devm_regmap_init(dev, NULL, client,
1630 + priv->chip->regmap_config);
1631 +- if (IS_ERR(priv->regmap))
1632 +- return PTR_ERR(priv->regmap);
1633 ++ if (IS_ERR(priv->regmap)) {
1634 ++ ret = PTR_ERR(priv->regmap);
1635 ++ goto disable_regs;
1636 ++ }
1637 +
1638 + priv->pdn_gpio = devm_gpiod_get_optional(dev, "pdn", GPIOD_OUT_LOW);
1639 + if (IS_ERR(priv->pdn_gpio)) {
1640 +@@ -849,7 +851,7 @@ static int tas571x_i2c_probe(struct i2c_client *client,
1641 +
1642 + ret = regmap_write(priv->regmap, TAS571X_OSC_TRIM_REG, 0);
1643 + if (ret)
1644 +- return ret;
1645 ++ goto disable_regs;
1646 +
1647 + usleep_range(50000, 60000);
1648 +
1649 +@@ -865,12 +867,20 @@ static int tas571x_i2c_probe(struct i2c_client *client,
1650 + */
1651 + ret = regmap_update_bits(priv->regmap, TAS571X_MVOL_REG, 1, 0);
1652 + if (ret)
1653 +- return ret;
1654 ++ goto disable_regs;
1655 + }
1656 +
1657 +- return devm_snd_soc_register_component(&client->dev,
1658 ++ ret = devm_snd_soc_register_component(&client->dev,
1659 + &priv->component_driver,
1660 + &tas571x_dai, 1);
1661 ++ if (ret)
1662 ++ goto disable_regs;
1663 ++
1664 ++ return ret;
1665 ++
1666 ++disable_regs:
1667 ++ regulator_bulk_disable(priv->chip->num_supply_names, priv->supplies);
1668 ++ return ret;
1669 + }
1670 +
1671 + static int tas571x_i2c_remove(struct i2c_client *client)
1672 +diff --git a/sound/soc/codecs/wm8960.c b/sound/soc/codecs/wm8960.c
1673 +index 8dc1f3d6a988..c4c00297ada6 100644
1674 +--- a/sound/soc/codecs/wm8960.c
1675 ++++ b/sound/soc/codecs/wm8960.c
1676 +@@ -863,8 +863,7 @@ static int wm8960_hw_params(struct snd_pcm_substream *substream,
1677 +
1678 + wm8960->is_stream_in_use[tx] = true;
1679 +
1680 +- if (snd_soc_component_get_bias_level(component) == SND_SOC_BIAS_ON &&
1681 +- !wm8960->is_stream_in_use[!tx])
1682 ++ if (!wm8960->is_stream_in_use[!tx])
1683 + return wm8960_configure_clocking(component);
1684 +
1685 + return 0;
1686 +diff --git a/sound/soc/qcom/qdsp6/q6afe-dai.c b/sound/soc/qcom/qdsp6/q6afe-dai.c
1687 +index 8f6c8fc073a9..1fc1939b90c2 100644
1688 +--- a/sound/soc/qcom/qdsp6/q6afe-dai.c
1689 ++++ b/sound/soc/qcom/qdsp6/q6afe-dai.c
1690 +@@ -899,6 +899,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1691 + SNDRV_PCM_RATE_16000,
1692 + .formats = SNDRV_PCM_FMTBIT_S16_LE |
1693 + SNDRV_PCM_FMTBIT_S24_LE,
1694 ++ .channels_min = 1,
1695 ++ .channels_max = 8,
1696 + .rate_min = 8000,
1697 + .rate_max = 48000,
1698 + },
1699 +@@ -914,6 +916,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1700 + SNDRV_PCM_RATE_16000,
1701 + .formats = SNDRV_PCM_FMTBIT_S16_LE |
1702 + SNDRV_PCM_FMTBIT_S24_LE,
1703 ++ .channels_min = 1,
1704 ++ .channels_max = 8,
1705 + .rate_min = 8000,
1706 + .rate_max = 48000,
1707 + },
1708 +@@ -928,6 +932,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1709 + .rates = SNDRV_PCM_RATE_48000 | SNDRV_PCM_RATE_8000 |
1710 + SNDRV_PCM_RATE_16000,
1711 + .formats = SNDRV_PCM_FMTBIT_S16_LE,
1712 ++ .channels_min = 1,
1713 ++ .channels_max = 8,
1714 + .rate_min = 8000,
1715 + .rate_max = 48000,
1716 + },
1717 +@@ -943,6 +949,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1718 + SNDRV_PCM_RATE_16000,
1719 + .formats = SNDRV_PCM_FMTBIT_S16_LE |
1720 + SNDRV_PCM_FMTBIT_S24_LE,
1721 ++ .channels_min = 1,
1722 ++ .channels_max = 8,
1723 + .rate_min = 8000,
1724 + .rate_max = 48000,
1725 + },
1726 +@@ -957,6 +965,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1727 + .rates = SNDRV_PCM_RATE_48000 | SNDRV_PCM_RATE_8000 |
1728 + SNDRV_PCM_RATE_16000,
1729 + .formats = SNDRV_PCM_FMTBIT_S16_LE,
1730 ++ .channels_min = 1,
1731 ++ .channels_max = 8,
1732 + .rate_min = 8000,
1733 + .rate_max = 48000,
1734 + },
1735 +@@ -972,6 +982,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1736 + SNDRV_PCM_RATE_16000,
1737 + .formats = SNDRV_PCM_FMTBIT_S16_LE |
1738 + SNDRV_PCM_FMTBIT_S24_LE,
1739 ++ .channels_min = 1,
1740 ++ .channels_max = 8,
1741 + .rate_min = 8000,
1742 + .rate_max = 48000,
1743 + },
1744 +@@ -986,6 +998,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1745 + .rates = SNDRV_PCM_RATE_48000 | SNDRV_PCM_RATE_8000 |
1746 + SNDRV_PCM_RATE_16000,
1747 + .formats = SNDRV_PCM_FMTBIT_S16_LE,
1748 ++ .channels_min = 1,
1749 ++ .channels_max = 8,
1750 + .rate_min = 8000,
1751 + .rate_max = 48000,
1752 + },
1753 +@@ -1001,6 +1015,8 @@ static struct snd_soc_dai_driver q6afe_dais[] = {
1754 + SNDRV_PCM_RATE_16000,
1755 + .formats = SNDRV_PCM_FMTBIT_S16_LE |
1756 + SNDRV_PCM_FMTBIT_S24_LE,
1757 ++ .channels_min = 1,
1758 ++ .channels_max = 8,
1759 + .rate_min = 8000,
1760 + .rate_max = 48000,
1761 + },
1762 +diff --git a/tools/objtool/check.c b/tools/objtool/check.c
1763 +index 9479c74af9ba..4613d796492a 100644
1764 +--- a/tools/objtool/check.c
1765 ++++ b/tools/objtool/check.c
1766 +@@ -2086,14 +2086,27 @@ static bool ignore_unreachable_insn(struct instruction *insn)
1767 + !strcmp(insn->sec->name, ".altinstr_aux"))
1768 + return true;
1769 +
1770 ++ if (!insn->func)
1771 ++ return false;
1772 ++
1773 ++ /*
1774 ++ * CONFIG_UBSAN_TRAP inserts a UD2 when it sees
1775 ++ * __builtin_unreachable(). The BUG() macro has an unreachable() after
1776 ++ * the UD2, which causes GCC's undefined trap logic to emit another UD2
1777 ++ * (or occasionally a JMP to UD2).
1778 ++ */
1779 ++ if (list_prev_entry(insn, list)->dead_end &&
1780 ++ (insn->type == INSN_BUG ||
1781 ++ (insn->type == INSN_JUMP_UNCONDITIONAL &&
1782 ++ insn->jump_dest && insn->jump_dest->type == INSN_BUG)))
1783 ++ return true;
1784 ++
1785 + /*
1786 + * Check if this (or a subsequent) instruction is related to
1787 + * CONFIG_UBSAN or CONFIG_KASAN.
1788 + *
1789 + * End the search at 5 instructions to avoid going into the weeds.
1790 + */
1791 +- if (!insn->func)
1792 +- return false;
1793 + for (i = 0; i < 5; i++) {
1794 +
1795 + if (is_kasan_insn(insn) || is_ubsan_insn(insn))
1796 +diff --git a/tools/objtool/orc_dump.c b/tools/objtool/orc_dump.c
1797 +index faa444270ee3..1a3e774941f8 100644
1798 +--- a/tools/objtool/orc_dump.c
1799 ++++ b/tools/objtool/orc_dump.c
1800 +@@ -78,7 +78,7 @@ int orc_dump(const char *_objname)
1801 + char *name;
1802 + size_t nr_sections;
1803 + Elf64_Addr orc_ip_addr = 0;
1804 +- size_t shstrtab_idx;
1805 ++ size_t shstrtab_idx, strtab_idx = 0;
1806 + Elf *elf;
1807 + Elf_Scn *scn;
1808 + GElf_Shdr sh;
1809 +@@ -139,6 +139,8 @@ int orc_dump(const char *_objname)
1810 +
1811 + if (!strcmp(name, ".symtab")) {
1812 + symtab = data;
1813 ++ } else if (!strcmp(name, ".strtab")) {
1814 ++ strtab_idx = i;
1815 + } else if (!strcmp(name, ".orc_unwind")) {
1816 + orc = data->d_buf;
1817 + orc_size = sh.sh_size;
1818 +@@ -150,7 +152,7 @@ int orc_dump(const char *_objname)
1819 + }
1820 + }
1821 +
1822 +- if (!symtab || !orc || !orc_ip)
1823 ++ if (!symtab || !strtab_idx || !orc || !orc_ip)
1824 + return 0;
1825 +
1826 + if (orc_size % sizeof(*orc) != 0) {
1827 +@@ -171,21 +173,29 @@ int orc_dump(const char *_objname)
1828 + return -1;
1829 + }
1830 +
1831 +- scn = elf_getscn(elf, sym.st_shndx);
1832 +- if (!scn) {
1833 +- WARN_ELF("elf_getscn");
1834 +- return -1;
1835 +- }
1836 +-
1837 +- if (!gelf_getshdr(scn, &sh)) {
1838 +- WARN_ELF("gelf_getshdr");
1839 +- return -1;
1840 +- }
1841 +-
1842 +- name = elf_strptr(elf, shstrtab_idx, sh.sh_name);
1843 +- if (!name || !*name) {
1844 +- WARN_ELF("elf_strptr");
1845 +- return -1;
1846 ++ if (GELF_ST_TYPE(sym.st_info) == STT_SECTION) {
1847 ++ scn = elf_getscn(elf, sym.st_shndx);
1848 ++ if (!scn) {
1849 ++ WARN_ELF("elf_getscn");
1850 ++ return -1;
1851 ++ }
1852 ++
1853 ++ if (!gelf_getshdr(scn, &sh)) {
1854 ++ WARN_ELF("gelf_getshdr");
1855 ++ return -1;
1856 ++ }
1857 ++
1858 ++ name = elf_strptr(elf, shstrtab_idx, sh.sh_name);
1859 ++ if (!name) {
1860 ++ WARN_ELF("elf_strptr");
1861 ++ return -1;
1862 ++ }
1863 ++ } else {
1864 ++ name = elf_strptr(elf, strtab_idx, sym.st_name);
1865 ++ if (!name) {
1866 ++ WARN_ELF("elf_strptr");
1867 ++ return -1;
1868 ++ }
1869 + }
1870 +
1871 + printf("%s+%llx:", name, (unsigned long long)rela.r_addend);