Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

treewide: Remove uninitialized_var() usage

Using uninitialized_var() is dangerous as it papers over real bugs[1]
(or can in the future), and suppresses unrelated compiler warnings
(e.g. "unused variable"). If the compiler thinks it is uninitialized,
either simply initialize the variable or make compiler changes.

In preparation for removing[2] the[3] macro[4], remove all remaining
needless uses with the following script:

git grep '\buninitialized_var\b' | cut -d: -f1 | sort -u | \
xargs perl -pi -e \
's/\buninitialized_var\(([^\)]+)\)/\1/g;
s:\s*/\* (GCC be quiet|to make compiler happy) \*/$::g;'

drivers/video/fbdev/riva/riva_hw.c was manually tweaked to avoid
pathological white-space.

No outstanding warnings were found building allmodconfig with GCC 9.3.0
for x86_64, i386, arm64, arm, powerpc, powerpc64le, s390x, mips, sparc64,
alpha, and m68k.

[1] https://lore.kernel.org/lkml/20200603174714.192027-1-glider@google.com/
[2] https://lore.kernel.org/lkml/CA+55aFw+Vbj0i=1TGqCR5vQkCzWJ0QxK6CernOU6eedsudAixw@mail.gmail.com/
[3] https://lore.kernel.org/lkml/CA+55aFwgbgqhbp1fkxvRKEpzyR5J8n1vKT1VZdz9knmPuXhOeg@mail.gmail.com/
[4] https://lore.kernel.org/lkml/CA+55aFz2500WfbKXAx8s67wrm9=yVJu65TpLgN_ybYNv0VEOKA@mail.gmail.com/

Reviewed-by: Leon Romanovsky <leonro@mellanox.com> # drivers/infiniband and mlx4/mlx5
Acked-by: Jason Gunthorpe <jgg@mellanox.com> # IB
Acked-by: Kalle Valo <kvalo@codeaurora.org> # wireless drivers
Reviewed-by: Chao Yu <yuchao0@huawei.com> # erofs
Signed-off-by: Kees Cook <keescook@chromium.org>

+278 -279
+1 -1
arch/arm/mach-sa1100/assabet.c
··· 652 652 */ 653 653 static void __init get_assabet_scr(void) 654 654 { 655 - unsigned long uninitialized_var(scr), i; 655 + unsigned long scr, i; 656 656 657 657 GPDR |= 0x3fc; /* Configure GPIO 9:2 as outputs */ 658 658 GPSR = 0x3fc; /* Write 0xFF to GPIO 9:2 */
+1 -1
arch/arm/mm/alignment.c
··· 799 799 static int 800 800 do_alignment(unsigned long addr, unsigned int fsr, struct pt_regs *regs) 801 801 { 802 - union offset_union uninitialized_var(offset); 802 + union offset_union offset; 803 803 unsigned long instrptr; 804 804 int (*handler)(unsigned long addr, u32 instr, struct pt_regs *regs); 805 805 unsigned int type;
+1 -1
arch/ia64/kernel/process.c
··· 445 445 do_copy_task_regs (struct task_struct *task, struct unw_frame_info *info, void *arg) 446 446 { 447 447 unsigned long mask, sp, nat_bits = 0, ar_rnat, urbs_end, cfm; 448 - unsigned long uninitialized_var(ip); /* GCC be quiet */ 448 + unsigned long ip; 449 449 elf_greg_t *dst = arg; 450 450 struct pt_regs *pt; 451 451 char nat;
+1 -1
arch/ia64/mm/discontig.c
··· 180 180 void __init setup_per_cpu_areas(void) 181 181 { 182 182 struct pcpu_alloc_info *ai; 183 - struct pcpu_group_info *uninitialized_var(gi); 183 + struct pcpu_group_info *gi; 184 184 unsigned int *cpu_map; 185 185 void *base; 186 186 unsigned long base_offset;
+1 -1
arch/ia64/mm/tlb.c
··· 369 369 370 370 void ia64_tlb_init(void) 371 371 { 372 - ia64_ptce_info_t uninitialized_var(ptce_info); /* GCC be quiet */ 372 + ia64_ptce_info_t ptce_info; 373 373 u64 tr_pgbits; 374 374 long status; 375 375 pal_vm_info_1_u_t vm_info_1;
+1 -1
arch/mips/lib/dump_tlb.c
··· 79 79 unsigned int pagemask, guestctl1 = 0, c0, c1, i; 80 80 unsigned long asidmask = cpu_asid_mask(&current_cpu_data); 81 81 int asidwidth = DIV_ROUND_UP(ilog2(asidmask) + 1, 4); 82 - unsigned long uninitialized_var(s_mmid); 82 + unsigned long s_mmid; 83 83 #ifdef CONFIG_32BIT 84 84 bool xpa = cpu_has_xpa && (read_c0_pagegrain() & PG_ELPA); 85 85 int pwidth = xpa ? 11 : 8;
+1 -1
arch/mips/mm/init.c
··· 83 83 static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) 84 84 { 85 85 enum fixed_addresses idx; 86 - unsigned int uninitialized_var(old_mmid); 86 + unsigned int old_mmid; 87 87 unsigned long vaddr, flags, entrylo; 88 88 unsigned long old_ctx; 89 89 pte_t pte;
+3 -3
arch/mips/mm/tlb-r4k.c
··· 119 119 if (size <= (current_cpu_data.tlbsizeftlbsets ? 120 120 current_cpu_data.tlbsize / 8 : 121 121 current_cpu_data.tlbsize / 2)) { 122 - unsigned long old_entryhi, uninitialized_var(old_mmid); 122 + unsigned long old_entryhi, old_mmid; 123 123 int newpid = cpu_asid(cpu, mm); 124 124 125 125 old_entryhi = read_c0_entryhi(); ··· 213 213 int cpu = smp_processor_id(); 214 214 215 215 if (cpu_context(cpu, vma->vm_mm) != 0) { 216 - unsigned long uninitialized_var(old_mmid); 216 + unsigned long old_mmid; 217 217 unsigned long flags, old_entryhi; 218 218 int idx; 219 219 ··· 382 382 #ifdef CONFIG_XPA 383 383 panic("Broken for XPA kernels"); 384 384 #else 385 - unsigned int uninitialized_var(old_mmid); 385 + unsigned int old_mmid; 386 386 unsigned long flags; 387 387 unsigned long wired; 388 388 unsigned long old_pagemask;
+1 -1
arch/powerpc/kvm/book3s_64_mmu_radix.c
··· 33 33 gva_t eaddr, void *to, void *from, 34 34 unsigned long n) 35 35 { 36 - int uninitialized_var(old_pid), old_lpid; 36 + int old_pid, old_lpid; 37 37 unsigned long quadrant, ret = n; 38 38 bool is_load = !!to; 39 39
+1 -1
arch/powerpc/kvm/powerpc.c
··· 1110 1110 static void kvmppc_complete_mmio_load(struct kvm_vcpu *vcpu) 1111 1111 { 1112 1112 struct kvm_run *run = vcpu->run; 1113 - u64 uninitialized_var(gpr); 1113 + u64 gpr; 1114 1114 1115 1115 if (run->mmio.len > sizeof(gpr)) { 1116 1116 printk(KERN_ERR "bad MMIO length: %d\n", run->mmio.len);
+1 -1
arch/powerpc/platforms/52xx/mpc52xx_pic.c
··· 340 340 { 341 341 int l1irq; 342 342 int l2irq; 343 - struct irq_chip *uninitialized_var(irqchip); 343 + struct irq_chip *irqchip; 344 344 void *hndlr; 345 345 int type; 346 346 u32 reg;
+1 -1
arch/s390/kernel/smp.c
··· 146 146 147 147 static inline int pcpu_stopped(struct pcpu *pcpu) 148 148 { 149 - u32 uninitialized_var(status); 149 + u32 status; 150 150 151 151 if (__pcpu_sigp(pcpu->address, SIGP_SENSE, 152 152 0, &status) != SIGP_CC_STATUS_STORED)
+5 -5
arch/x86/kernel/quirks.c
··· 95 95 static void ich_force_enable_hpet(struct pci_dev *dev) 96 96 { 97 97 u32 val; 98 - u32 uninitialized_var(rcba); 98 + u32 rcba; 99 99 int err = 0; 100 100 101 101 if (hpet_address || force_hpet_address) ··· 185 185 static void old_ich_force_hpet_resume(void) 186 186 { 187 187 u32 val; 188 - u32 uninitialized_var(gen_cntl); 188 + u32 gen_cntl; 189 189 190 190 if (!force_hpet_address || !cached_dev) 191 191 return; ··· 207 207 static void old_ich_force_enable_hpet(struct pci_dev *dev) 208 208 { 209 209 u32 val; 210 - u32 uninitialized_var(gen_cntl); 210 + u32 gen_cntl; 211 211 212 212 if (hpet_address || force_hpet_address) 213 213 return; ··· 298 298 299 299 static void vt8237_force_enable_hpet(struct pci_dev *dev) 300 300 { 301 - u32 uninitialized_var(val); 301 + u32 val; 302 302 303 303 if (hpet_address || force_hpet_address) 304 304 return; ··· 429 429 430 430 static void nvidia_force_enable_hpet(struct pci_dev *dev) 431 431 { 432 - u32 uninitialized_var(val); 432 + u32 val; 433 433 434 434 if (hpet_address || force_hpet_address) 435 435 return;
+1 -1
arch/x86/kvm/mmu/mmu.c
··· 1986 1986 unsigned long data) 1987 1987 { 1988 1988 u64 *sptep; 1989 - struct rmap_iterator uninitialized_var(iter); 1989 + struct rmap_iterator iter; 1990 1990 int young = 0; 1991 1991 1992 1992 for_each_rmap_spte(rmap_head, &iter, sptep)
+1 -1
arch/x86/kvm/mmu/paging_tmpl.h
··· 314 314 { 315 315 int ret; 316 316 pt_element_t pte; 317 - pt_element_t __user *uninitialized_var(ptep_user); 317 + pt_element_t __user *ptep_user; 318 318 gfn_t table_gfn; 319 319 u64 pt_access, pte_access; 320 320 unsigned index, accessed_dirty, pte_pkey;
+1 -1
arch/x86/kvm/x86.c
··· 9927 9927 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) 9928 9928 { 9929 9929 int i, r; 9930 - unsigned long hva, uninitialized_var(old_npages); 9930 + unsigned long hva, old_npages; 9931 9931 struct kvm_memslots *slots = kvm_memslots(kvm); 9932 9932 struct kvm_memory_slot *slot; 9933 9933
+1 -1
block/blk-merge.c
··· 473 473 struct scatterlist *sglist, 474 474 struct scatterlist **sg) 475 475 { 476 - struct bio_vec uninitialized_var(bvec), bvprv = { NULL }; 476 + struct bio_vec bvec, bvprv = { NULL }; 477 477 struct bvec_iter iter; 478 478 int nsegs = 0; 479 479 bool new_bio = false;
+1 -1
drivers/acpi/acpi_pad.c
··· 88 88 cpumask_var_t tmp; 89 89 int cpu; 90 90 unsigned long min_weight = -1; 91 - unsigned long uninitialized_var(preferred_cpu); 91 + unsigned long preferred_cpu; 92 92 93 93 if (!alloc_cpumask_var(&tmp, GFP_KERNEL)) 94 94 return;
+1 -1
drivers/ata/libata-scsi.c
··· 93 93 struct ata_link *link; 94 94 struct ata_device *dev; 95 95 unsigned long now; 96 - unsigned int uninitialized_var(msecs); 96 + unsigned int msecs; 97 97 int rc = 0; 98 98 99 99 ap = ata_shost_to_port(sdev->host);
+1 -1
drivers/atm/zatm.c
··· 940 940 vcc->qos.txtp.max_pcr >= ATM_OC3_PCR); 941 941 if (unlimited && zatm_dev->ubr != -1) zatm_vcc->shaper = zatm_dev->ubr; 942 942 else { 943 - int uninitialized_var(pcr); 943 + int pcr; 944 944 945 945 if (unlimited) vcc->qos.txtp.max_sdu = ATM_MAX_AAL5_PDU; 946 946 if ((zatm_vcc->shaper = alloc_shaper(vcc->dev,&pcr,
+3 -3
drivers/block/drbd/drbd_nl.c
··· 3423 3423 { 3424 3424 struct nlattr *resource_filter; 3425 3425 struct drbd_resource *resource; 3426 - struct drbd_device *uninitialized_var(device); 3426 + struct drbd_device *device; 3427 3427 int minor, err, retcode; 3428 3428 struct drbd_genlmsghdr *dh; 3429 3429 struct device_info device_info; ··· 3512 3512 { 3513 3513 struct nlattr *resource_filter; 3514 3514 struct drbd_resource *resource = NULL, *next_resource; 3515 - struct drbd_connection *uninitialized_var(connection); 3515 + struct drbd_connection *connection; 3516 3516 int err = 0, retcode; 3517 3517 struct drbd_genlmsghdr *dh; 3518 3518 struct connection_info connection_info; ··· 3674 3674 { 3675 3675 struct nlattr *resource_filter; 3676 3676 struct drbd_resource *resource; 3677 - struct drbd_device *uninitialized_var(device); 3677 + struct drbd_device *device; 3678 3678 struct drbd_peer_device *peer_device = NULL; 3679 3679 int minor, err, retcode; 3680 3680 struct drbd_genlmsghdr *dh;
+1 -1
drivers/block/rbd.c
··· 1993 1993 struct rbd_device *rbd_dev = obj_req->img_request->rbd_dev; 1994 1994 struct ceph_osd_data *osd_data; 1995 1995 u64 objno; 1996 - u8 state, new_state, uninitialized_var(current_state); 1996 + u8 state, new_state, current_state; 1997 1997 bool has_current_state; 1998 1998 void *p; 1999 1999
+1 -1
drivers/clk/clk-gate.c
··· 56 56 { 57 57 struct clk_gate *gate = to_clk_gate(hw); 58 58 int set = gate->flags & CLK_GATE_SET_TO_DISABLE ? 1 : 0; 59 - unsigned long uninitialized_var(flags); 59 + unsigned long flags; 60 60 u32 reg; 61 61 62 62 set ^= enable;
+7 -7
drivers/firewire/ohci.c
··· 1099 1099 static int context_add_buffer(struct context *ctx) 1100 1100 { 1101 1101 struct descriptor_buffer *desc; 1102 - dma_addr_t uninitialized_var(bus_addr); 1102 + dma_addr_t bus_addr; 1103 1103 int offset; 1104 1104 1105 1105 /* ··· 1289 1289 struct fw_packet *packet) 1290 1290 { 1291 1291 struct fw_ohci *ohci = ctx->ohci; 1292 - dma_addr_t d_bus, uninitialized_var(payload_bus); 1292 + dma_addr_t d_bus, payload_bus; 1293 1293 struct driver_data *driver_data; 1294 1294 struct descriptor *d, *last; 1295 1295 __le32 *header; ··· 2445 2445 { 2446 2446 struct fw_ohci *ohci; 2447 2447 __be32 *next_config_rom; 2448 - dma_addr_t uninitialized_var(next_config_rom_bus); 2448 + dma_addr_t next_config_rom_bus; 2449 2449 2450 2450 ohci = fw_ohci(card); 2451 2451 ··· 2933 2933 int type, int channel, size_t header_size) 2934 2934 { 2935 2935 struct fw_ohci *ohci = fw_ohci(card); 2936 - struct iso_context *uninitialized_var(ctx); 2937 - descriptor_callback_t uninitialized_var(callback); 2938 - u64 *uninitialized_var(channels); 2939 - u32 *uninitialized_var(mask), uninitialized_var(regs); 2936 + struct iso_context *ctx; 2937 + descriptor_callback_t callback; 2938 + u64 *channels; 2939 + u32 *mask, regs; 2940 2940 int index, ret = -EBUSY; 2941 2941 2942 2942 spin_lock_irq(&ohci->lock);
+1 -1
drivers/gpu/drm/bridge/sil-sii8620.c
··· 986 986 987 987 static void sii8620_stop_video(struct sii8620 *ctx) 988 988 { 989 - u8 uninitialized_var(val); 989 + u8 val; 990 990 991 991 sii8620_write_seq_static(ctx, 992 992 REG_TPI_INTR_EN, 0,
+1 -1
drivers/gpu/drm/drm_edid.c
··· 3051 3051 const u8 empty[3] = { 0, 0, 0 }; 3052 3052 3053 3053 for (i = 0; i < 4; i++) { 3054 - int uninitialized_var(width), height; 3054 + int width, height; 3055 3055 cvt = &(timing->data.other_data.data.cvt[i]); 3056 3056 3057 3057 if (!memcmp(cvt->code, empty, 3))
+3 -3
drivers/gpu/drm/exynos/exynos_drm_dsi.c
··· 547 547 unsigned long best_freq = 0; 548 548 u32 min_delta = 0xffffffff; 549 549 u8 p_min, p_max; 550 - u8 _p, uninitialized_var(best_p); 551 - u16 _m, uninitialized_var(best_m); 552 - u8 _s, uninitialized_var(best_s); 550 + u8 _p, best_p; 551 + u16 _m, best_m; 552 + u8 _s, best_s; 553 553 554 554 p_min = DIV_ROUND_UP(fin, (12 * MHZ)); 555 555 p_max = fin / (6 * MHZ);
+1 -1
drivers/gpu/drm/i915/display/intel_fbc.c
··· 474 474 unsigned int size, unsigned int fb_cpp) 475 475 { 476 476 struct intel_fbc *fbc = &dev_priv->fbc; 477 - struct drm_mm_node *uninitialized_var(compressed_llb); 477 + struct drm_mm_node *compressed_llb; 478 478 int ret; 479 479 480 480 drm_WARN_ON(&dev_priv->drm,
+1 -1
drivers/gpu/drm/i915/gt/intel_lrc.c
··· 1103 1103 __unwind_incomplete_requests(struct intel_engine_cs *engine) 1104 1104 { 1105 1105 struct i915_request *rq, *rn, *active = NULL; 1106 - struct list_head *uninitialized_var(pl); 1106 + struct list_head *pl; 1107 1107 int prio = I915_PRIORITY_INVALID; 1108 1108 1109 1109 lockdep_assert_held(&engine->active.lock);
+1 -1
drivers/gpu/drm/i915/intel_uncore.c
··· 1991 1991 unsigned int slow_timeout_ms, 1992 1992 u32 *out_value) 1993 1993 { 1994 - u32 uninitialized_var(reg_value); 1994 + u32 reg_value; 1995 1995 #define done (((reg_value = intel_uncore_read_fw(uncore, reg)) & mask) == value) 1996 1996 int ret; 1997 1997
+2 -2
drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c
··· 517 517 unsigned long best_freq = 0; 518 518 unsigned long fvco_min, fvco_max, fin, fout; 519 519 unsigned int min_prediv, max_prediv; 520 - unsigned int _prediv, uninitialized_var(best_prediv); 521 - unsigned long _fbdiv, uninitialized_var(best_fbdiv); 520 + unsigned int _prediv, best_prediv; 521 + unsigned long _fbdiv, best_fbdiv; 522 522 unsigned long min_delta = ULONG_MAX; 523 523 524 524 dsi->format = format;
+1 -1
drivers/i2c/busses/i2c-rk3x.c
··· 415 415 { 416 416 unsigned int i; 417 417 unsigned int len = i2c->msg->len - i2c->processed; 418 - u32 uninitialized_var(val); 418 + u32 val; 419 419 u8 byte; 420 420 421 421 /* we only care for MBRF here. */
+1 -1
drivers/ide/ide-acpi.c
··· 180 180 static acpi_handle ide_acpi_hwif_get_handle(ide_hwif_t *hwif) 181 181 { 182 182 struct device *dev = hwif->gendev.parent; 183 - acpi_handle uninitialized_var(dev_handle); 183 + acpi_handle dev_handle; 184 184 u64 pcidevfn; 185 185 acpi_handle chan_handle; 186 186 int err;
+1 -1
drivers/ide/ide-atapi.c
··· 609 609 610 610 static ide_startstop_t ide_transfer_pc(ide_drive_t *drive) 611 611 { 612 - struct ide_atapi_pc *uninitialized_var(pc); 612 + struct ide_atapi_pc *pc; 613 613 ide_hwif_t *hwif = drive->hwif; 614 614 struct request *rq = hwif->rq; 615 615 ide_expiry_t *expiry;
+2 -2
drivers/ide/ide-io-std.c
··· 173 173 u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; 174 174 175 175 if (io_32bit) { 176 - unsigned long uninitialized_var(flags); 176 + unsigned long flags; 177 177 178 178 if ((io_32bit & 2) && !mmio) { 179 179 local_irq_save(flags); ··· 217 217 u8 mmio = (hwif->host_flags & IDE_HFLAG_MMIO) ? 1 : 0; 218 218 219 219 if (io_32bit) { 220 - unsigned long uninitialized_var(flags); 220 + unsigned long flags; 221 221 222 222 if ((io_32bit & 2) && !mmio) { 223 223 local_irq_save(flags);
+4 -4
drivers/ide/ide-io.c
··· 622 622 void ide_timer_expiry (struct timer_list *t) 623 623 { 624 624 ide_hwif_t *hwif = from_timer(hwif, t, timer); 625 - ide_drive_t *uninitialized_var(drive); 625 + ide_drive_t *drive; 626 626 ide_handler_t *handler; 627 627 unsigned long flags; 628 628 int wait = -1; 629 629 int plug_device = 0; 630 - struct request *uninitialized_var(rq_in_flight); 630 + struct request *rq_in_flight; 631 631 632 632 spin_lock_irqsave(&hwif->lock, flags); 633 633 ··· 780 780 { 781 781 ide_hwif_t *hwif = (ide_hwif_t *)dev_id; 782 782 struct ide_host *host = hwif->host; 783 - ide_drive_t *uninitialized_var(drive); 783 + ide_drive_t *drive; 784 784 ide_handler_t *handler; 785 785 unsigned long flags; 786 786 ide_startstop_t startstop; 787 787 irqreturn_t irq_ret = IRQ_NONE; 788 788 int plug_device = 0; 789 - struct request *uninitialized_var(rq_in_flight); 789 + struct request *rq_in_flight; 790 790 791 791 if (host->host_flags & IDE_HFLAG_SERIALIZE) { 792 792 if (hwif != host->cur_port)
+1 -1
drivers/ide/ide-sysfs.c
··· 131 131 132 132 int ide_sysfs_register_port(ide_hwif_t *hwif) 133 133 { 134 - int i, uninitialized_var(rc); 134 + int i, rc; 135 135 136 136 for (i = 0; ide_port_attrs[i]; i++) { 137 137 rc = device_create_file(hwif->portdev, ide_port_attrs[i]);
+1 -1
drivers/ide/umc8672.c
··· 108 108 static void umc_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive) 109 109 { 110 110 ide_hwif_t *mate = hwif->mate; 111 - unsigned long uninitialized_var(flags); 111 + unsigned long flags; 112 112 const u8 pio = drive->pio_mode - XFER_PIO_0; 113 113 114 114 printk("%s: setting umc8672 to PIO mode%d (speed %d)\n",
+1 -1
drivers/idle/intel_idle.c
··· 132 132 struct cpuidle_state *state = &drv->states[index]; 133 133 unsigned long eax = flg2MWAIT(state->flags); 134 134 unsigned long ecx = 1; /* break on interrupt flag */ 135 - bool uninitialized_var(tick); 135 + bool tick; 136 136 int cpu = smp_processor_id(); 137 137 138 138 /*
+2 -2
drivers/infiniband/core/uverbs_cmd.c
··· 1584 1584 struct ib_uverbs_create_qp_resp resp; 1585 1585 struct ib_uqp_object *obj; 1586 1586 struct ib_xrcd *xrcd; 1587 - struct ib_uobject *uninitialized_var(xrcd_uobj); 1587 + struct ib_uobject *xrcd_uobj; 1588 1588 struct ib_qp *qp; 1589 1589 struct ib_qp_open_attr attr = {}; 1590 1590 int ret; ··· 3406 3406 struct ib_usrq_object *obj; 3407 3407 struct ib_pd *pd; 3408 3408 struct ib_srq *srq; 3409 - struct ib_uobject *uninitialized_var(xrcd_uobj); 3409 + struct ib_uobject *xrcd_uobj; 3410 3410 struct ib_srq_init_attr attr; 3411 3411 int ret; 3412 3412 struct ib_device *ib_dev;
+1 -1
drivers/infiniband/hw/cxgb4/cm.c
··· 3282 3282 3283 3283 static int pick_local_ip6addrs(struct c4iw_dev *dev, struct iw_cm_id *cm_id) 3284 3284 { 3285 - struct in6_addr uninitialized_var(addr); 3285 + struct in6_addr addr; 3286 3286 struct sockaddr_in6 *la6 = (struct sockaddr_in6 *)&cm_id->m_local_addr; 3287 3287 struct sockaddr_in6 *ra6 = (struct sockaddr_in6 *)&cm_id->m_remote_addr; 3288 3288
+1 -1
drivers/infiniband/hw/cxgb4/cq.c
··· 754 754 static int __c4iw_poll_cq_one(struct c4iw_cq *chp, struct c4iw_qp *qhp, 755 755 struct ib_wc *wc, struct c4iw_srq *srq) 756 756 { 757 - struct t4_cqe uninitialized_var(cqe); 757 + struct t4_cqe cqe; 758 758 struct t4_wq *wq = qhp ? &qhp->wq : NULL; 759 759 u32 credit = 0; 760 760 u8 cqe_flushed;
+3 -3
drivers/infiniband/hw/mlx4/qp.c
··· 3541 3541 int nreq; 3542 3542 int err = 0; 3543 3543 unsigned ind; 3544 - int uninitialized_var(size); 3545 - unsigned uninitialized_var(seglen); 3544 + int size; 3545 + unsigned seglen; 3546 3546 __be32 dummy; 3547 3547 __be32 *lso_wqe; 3548 - __be32 uninitialized_var(lso_hdr_sz); 3548 + __be32 lso_hdr_sz; 3549 3549 __be32 blh; 3550 3550 int i; 3551 3551 struct mlx4_ib_dev *mdev = to_mdev(ibqp->device);
+3 -3
drivers/infiniband/hw/mlx5/cq.c
··· 925 925 struct mlx5_ib_dev *dev = to_mdev(ibdev); 926 926 struct mlx5_ib_cq *cq = to_mcq(ibcq); 927 927 u32 out[MLX5_ST_SZ_DW(create_cq_out)]; 928 - int uninitialized_var(index); 929 - int uninitialized_var(inlen); 928 + int index; 929 + int inlen; 930 930 u32 *cqb = NULL; 931 931 void *cqc; 932 932 int cqe_size; ··· 1246 1246 __be64 *pas; 1247 1247 int page_shift; 1248 1248 int inlen; 1249 - int uninitialized_var(cqe_size); 1249 + int cqe_size; 1250 1250 unsigned long flags; 1251 1251 1252 1252 if (!MLX5_CAP_GEN(dev->mdev, cq_resize)) {
+1 -1
drivers/infiniband/hw/mlx5/devx.c
··· 2536 2536 { 2537 2537 struct devx_async_event_file *ev_file = filp->private_data; 2538 2538 struct devx_event_subscription *event_sub; 2539 - struct devx_async_event_data *uninitialized_var(event); 2539 + struct devx_async_event_data *event; 2540 2540 int ret = 0; 2541 2541 size_t eventsz; 2542 2542 bool omit_data;
+1 -1
drivers/infiniband/hw/mlx5/wr.c
··· 1249 1249 struct mlx5_wqe_xrc_seg *xrc; 1250 1250 struct mlx5_bf *bf; 1251 1251 void *cur_edge; 1252 - int uninitialized_var(size); 1252 + int size; 1253 1253 unsigned long flags; 1254 1254 unsigned int idx; 1255 1255 int err = 0;
+5 -5
drivers/infiniband/hw/mthca/mthca_qp.c
··· 1639 1639 * without initializing f0 and size0, and they are in fact 1640 1640 * never used uninitialized. 1641 1641 */ 1642 - int uninitialized_var(size0); 1643 - u32 uninitialized_var(f0); 1642 + int size0; 1643 + u32 f0; 1644 1644 int ind; 1645 1645 u8 op0 = 0; 1646 1646 ··· 1835 1835 * without initializing size0, and it is in fact never used 1836 1836 * uninitialized. 1837 1837 */ 1838 - int uninitialized_var(size0); 1838 + int size0; 1839 1839 int ind; 1840 1840 void *wqe; 1841 1841 void *prev_wqe; ··· 1943 1943 * without initializing f0 and size0, and they are in fact 1944 1944 * never used uninitialized. 1945 1945 */ 1946 - int uninitialized_var(size0); 1947 - u32 uninitialized_var(f0); 1946 + int size0; 1947 + u32 f0; 1948 1948 int ind; 1949 1949 u8 op0 = 0; 1950 1950
+1 -1
drivers/infiniband/sw/siw/siw_qp_rx.c
··· 333 333 struct siw_srq *srq; 334 334 struct siw_wqe *wqe = NULL; 335 335 bool srq_event = false; 336 - unsigned long uninitialized_var(flags); 336 + unsigned long flags; 337 337 338 338 srq = qp->srq; 339 339 if (srq) {
+1 -1
drivers/input/serio/serio_raw.c
··· 159 159 { 160 160 struct serio_raw_client *client = file->private_data; 161 161 struct serio_raw *serio_raw = client->serio_raw; 162 - char uninitialized_var(c); 162 + char c; 163 163 ssize_t read = 0; 164 164 int error; 165 165
+1 -1
drivers/iommu/intel/iommu.c
··· 2236 2236 unsigned long nr_pages, int prot) 2237 2237 { 2238 2238 struct dma_pte *first_pte = NULL, *pte = NULL; 2239 - phys_addr_t uninitialized_var(pteval); 2239 + phys_addr_t pteval; 2240 2240 unsigned long sg_res = 0; 2241 2241 unsigned int largepage_lvl = 0; 2242 2242 unsigned long lvl_pages = 0;
+1 -1
drivers/md/dm-io.c
··· 306 306 struct request_queue *q = bdev_get_queue(where->bdev); 307 307 unsigned short logical_block_size = queue_logical_block_size(q); 308 308 sector_t num_sectors; 309 - unsigned int uninitialized_var(special_cmd_max_sectors); 309 + unsigned int special_cmd_max_sectors; 310 310 311 311 /* 312 312 * Reject unsupported discard and write same requests.
+1 -1
drivers/md/dm-ioctl.c
··· 1844 1844 int ioctl_flags; 1845 1845 int param_flags; 1846 1846 unsigned int cmd; 1847 - struct dm_ioctl *uninitialized_var(param); 1847 + struct dm_ioctl *param; 1848 1848 ioctl_fn fn = NULL; 1849 1849 size_t input_param_size; 1850 1850 struct dm_ioctl param_kernel;
+1 -1
drivers/md/dm-snap-persistent.c
··· 613 613 chunk_t old, chunk_t new), 614 614 void *callback_context) 615 615 { 616 - int r, uninitialized_var(new_snapshot); 616 + int r, new_snapshot; 617 617 struct pstore *ps = get_info(store); 618 618 619 619 /*
+1 -1
drivers/md/dm-table.c
··· 642 642 */ 643 643 unsigned short remaining = 0; 644 644 645 - struct dm_target *uninitialized_var(ti); 645 + struct dm_target *ti; 646 646 struct queue_limits ti_limits; 647 647 unsigned i; 648 648
+1 -1
drivers/md/dm-writecache.c
··· 1752 1752 { 1753 1753 struct dm_writecache *wc = container_of(work, struct dm_writecache, writeback_work); 1754 1754 struct blk_plug plug; 1755 - struct wc_entry *f, *uninitialized_var(g), *e = NULL; 1755 + struct wc_entry *f, *g, *e = NULL; 1756 1756 struct rb_node *node, *next_node; 1757 1757 struct list_head skipped; 1758 1758 struct writeback_list wbl;
+1 -1
drivers/md/raid5.c
··· 2601 2601 struct stripe_head *sh = bi->bi_private; 2602 2602 struct r5conf *conf = sh->raid_conf; 2603 2603 int disks = sh->disks, i; 2604 - struct md_rdev *uninitialized_var(rdev); 2604 + struct md_rdev *rdev; 2605 2605 sector_t first_bad; 2606 2606 int bad_sectors; 2607 2607 int replacement = 0;
+1 -1
drivers/media/dvb-frontends/rtl2832.c
··· 640 640 struct i2c_client *client = dev->client; 641 641 struct dtv_frontend_properties *c = &fe->dtv_property_cache; 642 642 int ret; 643 - u32 uninitialized_var(tmp); 643 + u32 tmp; 644 644 u8 u8tmp, buf[2]; 645 645 u16 u16tmp; 646 646
+2 -2
drivers/media/tuners/qt1010.c
··· 215 215 static int qt1010_init_meas1(struct qt1010_priv *priv, 216 216 u8 oper, u8 reg, u8 reg_init_val, u8 *retval) 217 217 { 218 - u8 i, val1, uninitialized_var(val2); 218 + u8 i, val1, val2; 219 219 int err; 220 220 221 221 qt1010_i2c_oper_t i2c_data[] = { ··· 250 250 static int qt1010_init_meas2(struct qt1010_priv *priv, 251 251 u8 reg_init_val, u8 *retval) 252 252 { 253 - u8 i, uninitialized_var(val); 253 + u8 i, val; 254 254 int err; 255 255 qt1010_i2c_oper_t i2c_data[] = { 256 256 { QT1010_WR, 0x07, reg_init_val },
+1 -1
drivers/media/usb/gspca/vicam.c
··· 225 225 { 226 226 int ret; 227 227 const struct ihex_binrec *rec; 228 - const struct firmware *uninitialized_var(fw); 228 + const struct firmware *fw; 229 229 u8 *firmware_buf; 230 230 231 231 ret = request_ihex_firmware(&fw, VICAM_FIRMWARE,
+4 -4
drivers/media/usb/uvc/uvc_video.c
··· 765 765 unsigned int header_size; 766 766 bool has_pts = false; 767 767 bool has_scr = false; 768 - u16 uninitialized_var(scr_sof); 769 - u32 uninitialized_var(scr_stc); 770 - u32 uninitialized_var(pts); 768 + u16 scr_sof; 769 + u32 scr_stc; 770 + u32 pts; 771 771 772 772 if (stream->stats.stream.nb_frames == 0 && 773 773 stream->stats.frame.nb_packets == 0) ··· 1828 1828 struct usb_host_endpoint *best_ep = NULL; 1829 1829 unsigned int best_psize = UINT_MAX; 1830 1830 unsigned int bandwidth; 1831 - unsigned int uninitialized_var(altsetting); 1831 + unsigned int altsetting; 1832 1832 int intfnum = stream->intfnum; 1833 1833 1834 1834 /* Isochronous endpoint, select the alternate setting. */
+1 -1
drivers/memstick/host/jmb38x_ms.c
··· 314 314 } 315 315 316 316 while (length) { 317 - unsigned int uninitialized_var(p_off); 317 + unsigned int p_off; 318 318 319 319 if (host->req->long_data) { 320 320 pg = nth_page(sg_page(&host->req->sg),
+1 -1
drivers/memstick/host/tifm_ms.c
··· 198 198 host->block_pos); 199 199 200 200 while (length) { 201 - unsigned int uninitialized_var(p_off); 201 + unsigned int p_off; 202 202 203 203 if (host->req->long_data) { 204 204 pg = nth_page(sg_page(&host->req->sg),
+1 -1
drivers/mmc/host/sdhci.c
··· 493 493 { 494 494 unsigned long flags; 495 495 size_t blksize, len, chunk; 496 - u32 uninitialized_var(scratch); 496 + u32 scratch; 497 497 u8 *buf; 498 498 499 499 DBG("PIO reading\n");
+1 -1
drivers/mtd/nand/raw/nand_ecc.c
··· 131 131 /* rp0..rp15..rp17 are the various accumulated parities (per byte) */ 132 132 uint32_t rp0, rp1, rp2, rp3, rp4, rp5, rp6, rp7; 133 133 uint32_t rp8, rp9, rp10, rp11, rp12, rp13, rp14, rp15, rp16; 134 - uint32_t uninitialized_var(rp17); /* to make compiler happy */ 134 + uint32_t rp17; 135 135 uint32_t par; /* the cumulative parity for all data */ 136 136 uint32_t tmppar; /* the cumulative parity for this iteration; 137 137 for rp12, rp14 and rp16 at the end of the
+1 -1
drivers/mtd/nand/raw/s3c2410.c
··· 291 291 int tacls_max = (info->cpu_type == TYPE_S3C2412) ? 8 : 4; 292 292 int tacls, twrph0, twrph1; 293 293 unsigned long clkrate = clk_get_rate(info->clk); 294 - unsigned long uninitialized_var(set), cfg, uninitialized_var(mask); 294 + unsigned long set, cfg, mask; 295 295 unsigned long flags; 296 296 297 297 /* calculate the timing information for the controller */
+2 -2
drivers/mtd/parsers/afs.c
··· 126 126 * Static checks cannot see that we bail out if we have an error 127 127 * reading the footer. 128 128 */ 129 - u_int uninitialized_var(iis_ptr); 130 - u_int uninitialized_var(img_ptr); 129 + u_int iis_ptr; 130 + u_int img_ptr; 131 131 u_int ptr; 132 132 size_t sz; 133 133 int ret;
+1 -1
drivers/mtd/ubi/eba.c
··· 599 599 int err, pnum, scrub = 0, vol_id = vol->vol_id; 600 600 struct ubi_vid_io_buf *vidb; 601 601 struct ubi_vid_hdr *vid_hdr; 602 - uint32_t uninitialized_var(crc); 602 + uint32_t crc; 603 603 604 604 err = leb_read_lock(ubi, vol_id, lnum); 605 605 if (err)
+1 -1
drivers/net/can/janz-ican3.c
··· 1451 1451 1452 1452 /* process all communication messages */ 1453 1453 while (true) { 1454 - struct ican3_msg uninitialized_var(msg); 1454 + struct ican3_msg msg; 1455 1455 ret = ican3_recv_msg(mod, &msg); 1456 1456 if (ret) 1457 1457 break;
+2 -2
drivers/net/ethernet/broadcom/bnx2.c
··· 1455 1455 static void 1456 1456 bnx2_enable_forced_2g5(struct bnx2 *bp) 1457 1457 { 1458 - u32 uninitialized_var(bmcr); 1458 + u32 bmcr; 1459 1459 int err; 1460 1460 1461 1461 if (!(bp->phy_flags & BNX2_PHY_FLAG_2_5G_CAPABLE)) ··· 1499 1499 static void 1500 1500 bnx2_disable_forced_2g5(struct bnx2 *bp) 1501 1501 { 1502 - u32 uninitialized_var(bmcr); 1502 + u32 bmcr; 1503 1503 int err; 1504 1504 1505 1505 if (!(bp->phy_flags & BNX2_PHY_FLAG_2_5G_CAPABLE))
+2 -2
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
··· 529 529 530 530 int mlx5_satisfy_startup_pages(struct mlx5_core_dev *dev, int boot) 531 531 { 532 - u16 uninitialized_var(func_id); 533 - s32 uninitialized_var(npages); 532 + u16 func_id; 533 + s32 npages; 534 534 int err; 535 535 536 536 err = mlx5_cmd_query_pages(dev, &func_id, &npages, boot);
+1 -1
drivers/net/ethernet/neterion/s2io.c
··· 7276 7276 int ring_no = ring_data->ring_no; 7277 7277 u16 l3_csum, l4_csum; 7278 7278 unsigned long long err = rxdp->Control_1 & RXD_T_CODE; 7279 - struct lro *uninitialized_var(lro); 7279 + struct lro *lro; 7280 7280 u8 err_mask; 7281 7281 struct swStat *swstats = &sp->mac_control.stats_info->sw_stat; 7282 7282
+1 -1
drivers/net/ethernet/qlogic/qla3xxx.c
··· 3769 3769 struct net_device *ndev = NULL; 3770 3770 struct ql3_adapter *qdev = NULL; 3771 3771 static int cards_found; 3772 - int uninitialized_var(pci_using_dac), err; 3772 + int pci_using_dac, err; 3773 3773 3774 3774 err = pci_enable_device(pdev); 3775 3775 if (err) {
+1 -1
drivers/net/ethernet/sun/cassini.c
··· 2271 2271 drops = 0; 2272 2272 while (1) { 2273 2273 struct cas_rx_comp *rxc = rxcs + entry; 2274 - struct sk_buff *uninitialized_var(skb); 2274 + struct sk_buff *skb; 2275 2275 int type, len; 2276 2276 u64 words[4]; 2277 2277 int i, dring;
+3 -3
drivers/net/ethernet/sun/niu.c
··· 429 429 struct niu_link_config *lp = &np->link_config; 430 430 u16 pll_cfg, pll_sts; 431 431 int max_retry = 100; 432 - u64 uninitialized_var(sig), mask, val; 432 + u64 sig, mask, val; 433 433 u32 tx_cfg, rx_cfg; 434 434 unsigned long i; 435 435 int err; ··· 526 526 struct niu_link_config *lp = &np->link_config; 527 527 u32 tx_cfg, rx_cfg, pll_cfg, pll_sts; 528 528 int max_retry = 100; 529 - u64 uninitialized_var(sig), mask, val; 529 + u64 sig, mask, val; 530 530 unsigned long i; 531 531 int err; 532 532 ··· 714 714 715 715 static int esr_reset(struct niu *np) 716 716 { 717 - u32 uninitialized_var(reset); 717 + u32 reset; 718 718 int err; 719 719 720 720 err = mdio_write(np, np->port, NIU_ESR_DEV_ADDR,
+1 -1
drivers/net/wan/z85230.c
··· 702 702 irqreturn_t z8530_interrupt(int irq, void *dev_id) 703 703 { 704 704 struct z8530_dev *dev=dev_id; 705 - u8 uninitialized_var(intr); 705 + u8 intr; 706 706 static volatile int locker=0; 707 707 int work=0; 708 708 struct z8530_irqhandler *irqs;
+1 -1
drivers/net/wireless/ath/ath10k/core.c
··· 2240 2240 2241 2241 static int ath10k_init_hw_params(struct ath10k *ar) 2242 2242 { 2243 - const struct ath10k_hw_params *uninitialized_var(hw_params); 2243 + const struct ath10k_hw_params *hw_params; 2244 2244 int i; 2245 2245 2246 2246 for (i = 0; i < ARRAY_SIZE(ath10k_hw_params_list); i++) {
+1 -1
drivers/net/wireless/ath/ath6kl/init.c
··· 1575 1575 1576 1576 int ath6kl_init_hw_params(struct ath6kl *ar) 1577 1577 { 1578 - const struct ath6kl_hw *uninitialized_var(hw); 1578 + const struct ath6kl_hw *hw; 1579 1579 int i; 1580 1580 1581 1581 for (i = 0; i < ARRAY_SIZE(hw_list); i++) {
+1 -1
drivers/net/wireless/ath/ath9k/init.c
··· 230 230 struct ath_hw *ah = hw_priv; 231 231 struct ath_common *common = ath9k_hw_common(ah); 232 232 struct ath_softc *sc = (struct ath_softc *) common->priv; 233 - unsigned long uninitialized_var(flags); 233 + unsigned long flags; 234 234 u32 val; 235 235 236 236 if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_ON) {
+1 -1
drivers/net/wireless/broadcom/b43/debugfs.c
··· 493 493 struct b43_wldev *dev; 494 494 struct b43_debugfs_fops *dfops; 495 495 struct b43_dfs_file *dfile; 496 - ssize_t uninitialized_var(ret); 496 + ssize_t ret; 497 497 char *buf; 498 498 const size_t bufsize = 1024 * 16; /* 16 kiB buffer */ 499 499 const size_t buforder = get_order(bufsize);
+1 -1
drivers/net/wireless/broadcom/b43/dma.c
··· 37 37 static u32 b43_dma_address(struct b43_dma *dma, dma_addr_t dmaaddr, 38 38 enum b43_addrtype addrtype) 39 39 { 40 - u32 uninitialized_var(addr); 40 + u32 addr; 41 41 42 42 switch (addrtype) { 43 43 case B43_DMA_ADDR_LOW:
+1 -1
drivers/net/wireless/broadcom/b43/lo.c
··· 729 729 }; 730 730 int max_rx_gain; 731 731 struct b43_lo_calib *cal; 732 - struct lo_g_saved_values uninitialized_var(saved_regs); 732 + struct lo_g_saved_values saved_regs; 733 733 /* Values from the "TXCTL Register and Value Table" */ 734 734 u16 txctl_reg; 735 735 u16 txctl_value;
+1 -1
drivers/net/wireless/broadcom/b43/phy_n.c
··· 5643 5643 u8 rfctl[2]; 5644 5644 u8 afectl_core; 5645 5645 u16 tmp[6]; 5646 - u16 uninitialized_var(cur_hpf1), uninitialized_var(cur_hpf2), cur_lna; 5646 + u16 cur_hpf1, cur_hpf2, cur_lna; 5647 5647 u32 real, imag; 5648 5648 enum nl80211_band band; 5649 5649
+6 -6
drivers/net/wireless/broadcom/b43/xmit.c
··· 422 422 if ((rates[0].flags & IEEE80211_TX_RC_USE_RTS_CTS) || 423 423 (rates[0].flags & IEEE80211_TX_RC_USE_CTS_PROTECT)) { 424 424 unsigned int len; 425 - struct ieee80211_hdr *uninitialized_var(hdr); 425 + struct ieee80211_hdr *hdr; 426 426 int rts_rate, rts_rate_fb; 427 427 int rts_rate_ofdm, rts_rate_fb_ofdm; 428 - struct b43_plcp_hdr6 *uninitialized_var(plcp); 428 + struct b43_plcp_hdr6 *plcp; 429 429 struct ieee80211_rate *rts_cts_rate; 430 430 431 431 rts_cts_rate = ieee80211_get_rts_cts_rate(dev->wl->hw, info); ··· 436 436 rts_rate_fb_ofdm = b43_is_ofdm_rate(rts_rate_fb); 437 437 438 438 if (rates[0].flags & IEEE80211_TX_RC_USE_CTS_PROTECT) { 439 - struct ieee80211_cts *uninitialized_var(cts); 439 + struct ieee80211_cts *cts; 440 440 441 441 switch (dev->fw.hdr_format) { 442 442 case B43_FW_HDR_598: ··· 458 458 mac_ctl |= B43_TXH_MAC_SENDCTS; 459 459 len = sizeof(struct ieee80211_cts); 460 460 } else { 461 - struct ieee80211_rts *uninitialized_var(rts); 461 + struct ieee80211_rts *rts; 462 462 463 463 switch (dev->fw.hdr_format) { 464 464 case B43_FW_HDR_598: ··· 637 637 const struct b43_rxhdr_fw4 *rxhdr = _rxhdr; 638 638 __le16 fctl; 639 639 u16 phystat0, phystat3; 640 - u16 uninitialized_var(chanstat), uninitialized_var(mactime); 641 - u32 uninitialized_var(macstat); 640 + u16 chanstat, mactime; 641 + u32 macstat; 642 642 u16 chanid; 643 643 int padding, rate_idx; 644 644
+1 -1
drivers/net/wireless/broadcom/b43legacy/debugfs.c
··· 190 190 struct b43legacy_wldev *dev; 191 191 struct b43legacy_debugfs_fops *dfops; 192 192 struct b43legacy_dfs_file *dfile; 193 - ssize_t uninitialized_var(ret); 193 + ssize_t ret; 194 194 char *buf; 195 195 const size_t bufsize = 1024 * 16; /* 16 KiB buffer */ 196 196 const size_t buforder = get_order(bufsize);
+1 -1
drivers/net/wireless/broadcom/b43legacy/main.c
··· 2580 2580 static int b43legacy_switch_phymode(struct b43legacy_wl *wl, 2581 2581 unsigned int new_mode) 2582 2582 { 2583 - struct b43legacy_wldev *uninitialized_var(up_dev); 2583 + struct b43legacy_wldev *up_dev; 2584 2584 struct b43legacy_wldev *down_dev; 2585 2585 int err; 2586 2586 bool gmode = false;
+1 -1
drivers/net/wireless/intel/iwlegacy/3945.c
··· 2100 2100 2101 2101 /* set tx power value for all OFDM rates */ 2102 2102 for (rate_idx = 0; rate_idx < IL_OFDM_RATES; rate_idx++) { 2103 - s32 uninitialized_var(power_idx); 2103 + s32 power_idx; 2104 2104 int rc; 2105 2105 2106 2106 /* use channel group's clip-power table,
+1 -1
drivers/net/wireless/intel/iwlegacy/4965-mac.c
··· 2769 2769 struct ieee80211_tx_info *info; 2770 2770 struct il4965_tx_resp *tx_resp = (void *)&pkt->u.raw[0]; 2771 2771 u32 status = le32_to_cpu(tx_resp->u.status); 2772 - int uninitialized_var(tid); 2772 + int tid; 2773 2773 int sta_id; 2774 2774 int freed; 2775 2775 u8 *qc = NULL;
+2 -2
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c
··· 618 618 u8 queue_sel) 619 619 { 620 620 u16 beq, bkq, viq, voq, mgtq, hiq; 621 - u16 uninitialized_var(valuehi); 622 - u16 uninitialized_var(valuelow); 621 + u16 valuehi; 622 + u16 valuelow; 623 623 624 624 switch (queue_sel) { 625 625 case (TX_SELE_HQ | TX_SELE_LQ):
+1 -1
drivers/pci/pcie/aer.c
··· 1138 1138 { 1139 1139 struct pcie_device *dev = (struct pcie_device *)context; 1140 1140 struct aer_rpc *rpc = get_service_data(dev); 1141 - struct aer_err_source uninitialized_var(e_src); 1141 + struct aer_err_source e_src; 1142 1142 1143 1143 if (kfifo_is_empty(&rpc->aer_fifo)) 1144 1144 return IRQ_NONE;
+2 -2
drivers/platform/x86/hdaps.c
··· 365 365 static ssize_t hdaps_temp1_show(struct device *dev, 366 366 struct device_attribute *attr, char *buf) 367 367 { 368 - u8 uninitialized_var(temp); 368 + u8 temp; 369 369 int ret; 370 370 371 371 ret = hdaps_readb_one(HDAPS_PORT_TEMP1, &temp); ··· 378 378 static ssize_t hdaps_temp2_show(struct device *dev, 379 379 struct device_attribute *attr, char *buf) 380 380 { 381 - u8 uninitialized_var(temp); 381 + u8 temp; 382 382 int ret; 383 383 384 384 ret = hdaps_readb_one(HDAPS_PORT_TEMP2, &temp);
+1 -1
drivers/scsi/dc395x.c
··· 4126 4126 const unsigned srbs_per_page = PAGE_SIZE/SEGMENTX_LEN; 4127 4127 int srb_idx = 0; 4128 4128 unsigned i = 0; 4129 - struct SGentry *uninitialized_var(ptr); 4129 + struct SGentry *ptr; 4130 4130 4131 4131 for (i = 0; i < DC395x_MAX_SRB_CNT; i++) 4132 4132 acb->srb_array[i].segment_x = NULL;
+1 -1
drivers/scsi/pm8001/pm8001_hwi.c
··· 4202 4202 { 4203 4203 struct outbound_queue_table *circularQ; 4204 4204 void *pMsg1 = NULL; 4205 - u8 uninitialized_var(bc); 4205 + u8 bc; 4206 4206 u32 ret = MPI_IO_STATUS_FAIL; 4207 4207 unsigned long flags; 4208 4208
+1 -1
drivers/scsi/pm8001/pm80xx_hwi.c
··· 4182 4182 { 4183 4183 struct outbound_queue_table *circularQ; 4184 4184 void *pMsg1 = NULL; 4185 - u8 uninitialized_var(bc); 4185 + u8 bc; 4186 4186 u32 ret = MPI_IO_STATUS_FAIL; 4187 4187 unsigned long flags; 4188 4188 u32 regval;
+2 -2
drivers/ssb/driver_chipcommon.c
··· 119 119 static enum ssb_clksrc chipco_pctl_get_slowclksrc(struct ssb_chipcommon *cc) 120 120 { 121 121 struct ssb_bus *bus = cc->dev->bus; 122 - u32 uninitialized_var(tmp); 122 + u32 tmp; 123 123 124 124 if (cc->dev->id.revision < 6) { 125 125 if (bus->bustype == SSB_BUSTYPE_SSB || ··· 149 149 /* Get maximum or minimum (depending on get_max flag) slowclock frequency. */ 150 150 static int chipco_pctl_clockfreqlimit(struct ssb_chipcommon *cc, int get_max) 151 151 { 152 - int uninitialized_var(limit); 152 + int limit; 153 153 enum ssb_clksrc clocksrc; 154 154 int divisor = 1; 155 155 u32 tmp;
+1 -1
drivers/tty/cyclades.c
··· 3643 3643 struct cyclades_card *card; 3644 3644 void __iomem *addr0 = NULL, *addr2 = NULL; 3645 3645 char *card_name = NULL; 3646 - u32 uninitialized_var(mailbox); 3646 + u32 mailbox; 3647 3647 unsigned int device_id, nchan = 0, card_no, i, j; 3648 3648 unsigned char plx_ver; 3649 3649 int retval, irq;
+1 -1
drivers/tty/isicom.c
··· 1514 1514 static int isicom_probe(struct pci_dev *pdev, 1515 1515 const struct pci_device_id *ent) 1516 1516 { 1517 - unsigned int uninitialized_var(signature), index; 1517 + unsigned int signature, index; 1518 1518 int retval = -EPERM; 1519 1519 struct isi_board *board = NULL; 1520 1520
+1 -1
drivers/usb/musb/cppi_dma.c
··· 1146 1146 struct musb_hw_ep *hw_ep = NULL; 1147 1147 u32 rx, tx; 1148 1148 int i, index; 1149 - unsigned long uninitialized_var(flags); 1149 + unsigned long flags; 1150 1150 1151 1151 cppi = container_of(musb->dma_controller, struct cppi, controller); 1152 1152 if (cppi->irq)
+2 -2
drivers/usb/storage/sddr55.c
··· 554 554 555 555 static unsigned long sddr55_get_capacity(struct us_data *us) { 556 556 557 - unsigned char uninitialized_var(manufacturerID); 558 - unsigned char uninitialized_var(deviceID); 557 + unsigned char manufacturerID; 558 + unsigned char deviceID; 559 559 int result; 560 560 struct sddr55_card_info *info = (struct sddr55_card_info *)us->extra; 561 561
+3 -3
drivers/vhost/net.c
··· 862 862 struct tun_msg_ctl ctl; 863 863 size_t len, total_len = 0; 864 864 int err; 865 - struct vhost_net_ubuf_ref *uninitialized_var(ubufs); 865 + struct vhost_net_ubuf_ref *ubufs; 866 866 bool zcopy_used; 867 867 int sent_pkts = 0; 868 868 ··· 1042 1042 /* len is always initialized before use since we are always called with 1043 1043 * datalen > 0. 1044 1044 */ 1045 - u32 uninitialized_var(len); 1045 + u32 len; 1046 1046 1047 1047 while (datalen > 0 && headcount < quota) { 1048 1048 if (unlikely(seg >= UIO_MAXIOV)) { ··· 1099 1099 { 1100 1100 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_RX]; 1101 1101 struct vhost_virtqueue *vq = &nvq->vq; 1102 - unsigned uninitialized_var(in), log; 1102 + unsigned in, log; 1103 1103 struct vhost_log *vq_log; 1104 1104 struct msghdr msg = { 1105 1105 .msg_name = NULL,
+3 -3
drivers/video/fbdev/matrox/matroxfb_maven.c
··· 300 300 unsigned int* in, unsigned int* feed, unsigned int* post, 301 301 unsigned int* htotal2) { 302 302 unsigned int fvco; 303 - unsigned int uninitialized_var(p); 303 + unsigned int p; 304 304 305 305 fvco = matroxfb_PLL_mavenclock(&maven1000_pll, ctl, htotal, vtotal, in, feed, &p, htotal2); 306 306 if (!fvco) ··· 732 732 733 733 for (x = 0; x < 8; x++) { 734 734 unsigned int c; 735 - unsigned int uninitialized_var(a), uninitialized_var(b), 736 - uninitialized_var(h2); 735 + unsigned int a, b, 736 + h2; 737 737 unsigned int h = ht + 2 + x; 738 738 739 739 if (!matroxfb_mavenclock((m->mode == MATROXFB_OUTPUT_MODE_PAL) ? &maven_PAL : &maven_NTSC, h, vt, &a, &b, &c, &h2)) {
+3 -3
drivers/video/fbdev/pm3fb.c
··· 821 821 822 822 wmb(); 823 823 { 824 - unsigned char uninitialized_var(m); /* ClkPreScale */ 825 - unsigned char uninitialized_var(n); /* ClkFeedBackScale */ 826 - unsigned char uninitialized_var(p); /* ClkPostScale */ 824 + unsigned char m; /* ClkPreScale */ 825 + unsigned char n; /* ClkFeedBackScale */ 826 + unsigned char p; /* ClkPostScale */ 827 827 unsigned long pixclock = PICOS2KHZ(info->var.pixclock); 828 828 829 829 (void)pm3fb_calculate_clock(pixclock, &m, &n, &p);
+1 -2
drivers/video/fbdev/riva/riva_hw.c
··· 1245 1245 ) 1246 1246 { 1247 1247 int pixelDepth; 1248 - int uninitialized_var(VClk),uninitialized_var(m), 1249 - uninitialized_var(n), uninitialized_var(p); 1248 + int VClk, m, n, p; 1250 1249 1251 1250 /* 1252 1251 * Save mode parameters.
+3 -3
drivers/virtio/virtio_ring.c
··· 424 424 struct vring_virtqueue *vq = to_vvq(_vq); 425 425 struct scatterlist *sg; 426 426 struct vring_desc *desc; 427 - unsigned int i, n, avail, descs_used, uninitialized_var(prev), err_idx; 427 + unsigned int i, n, avail, descs_used, prev, err_idx; 428 428 int head; 429 429 bool indirect; 430 430 ··· 1101 1101 struct vring_packed_desc *desc; 1102 1102 struct scatterlist *sg; 1103 1103 unsigned int i, n, c, descs_used, err_idx; 1104 - __le16 uninitialized_var(head_flags), flags; 1105 - u16 head, id, uninitialized_var(prev), curr, avail_used_flags; 1104 + __le16 head_flags, flags; 1105 + u16 head, id, prev, curr, avail_used_flags; 1106 1106 1107 1107 START_USE(vq); 1108 1108
+1 -1
fs/afs/dir.c
··· 1053 1053 static int afs_d_revalidate(struct dentry *dentry, unsigned int flags) 1054 1054 { 1055 1055 struct afs_vnode *vnode, *dir; 1056 - struct afs_fid uninitialized_var(fid); 1056 + struct afs_fid fid; 1057 1057 struct dentry *parent; 1058 1058 struct inode *inode; 1059 1059 struct key *key;
+1 -1
fs/afs/security.c
··· 399 399 int afs_permission(struct inode *inode, int mask) 400 400 { 401 401 struct afs_vnode *vnode = AFS_FS_I(inode); 402 - afs_access_t uninitialized_var(access); 402 + afs_access_t access; 403 403 struct key *key; 404 404 int ret = 0; 405 405
+1 -1
fs/dlm/netlink.c
··· 113 113 114 114 void dlm_timeout_warn(struct dlm_lkb *lkb) 115 115 { 116 - struct sk_buff *uninitialized_var(send_skb); 116 + struct sk_buff *send_skb; 117 117 struct dlm_lock_data *data; 118 118 size_t size; 119 119 int rv;
+2 -2
fs/erofs/data.c
··· 265 265 */ 266 266 static int erofs_raw_access_readpage(struct file *file, struct page *page) 267 267 { 268 - erofs_off_t uninitialized_var(last_block); 268 + erofs_off_t last_block; 269 269 struct bio *bio; 270 270 271 271 trace_erofs_readpage(page, true); ··· 282 282 283 283 static void erofs_raw_access_readahead(struct readahead_control *rac) 284 284 { 285 - erofs_off_t uninitialized_var(last_block); 285 + erofs_off_t last_block; 286 286 struct bio *bio = NULL; 287 287 struct page *page; 288 288
+1 -1
fs/erofs/zdata.c
··· 1161 1161 struct z_erofs_decompressqueue *q[NR_JOBQUEUES]; 1162 1162 void *bi_private; 1163 1163 /* since bio will be NULL, no need to initialize last_index */ 1164 - pgoff_t uninitialized_var(last_index); 1164 + pgoff_t last_index; 1165 1165 unsigned int nr_bios = 0; 1166 1166 struct bio *bio = NULL; 1167 1167
+1 -1
fs/fat/dir.c
··· 1284 1284 struct super_block *sb = dir->i_sb; 1285 1285 struct msdos_sb_info *sbi = MSDOS_SB(sb); 1286 1286 struct buffer_head *bh, *prev, *bhs[3]; /* 32*slots (672bytes) */ 1287 - struct msdos_dir_entry *uninitialized_var(de); 1287 + struct msdos_dir_entry *de; 1288 1288 int err, free_slots, i, nr_bhs; 1289 1289 loff_t pos, i_pos; 1290 1290
+2 -2
fs/fuse/control.c
··· 120 120 const char __user *buf, 121 121 size_t count, loff_t *ppos) 122 122 { 123 - unsigned uninitialized_var(val); 123 + unsigned val; 124 124 ssize_t ret; 125 125 126 126 ret = fuse_conn_limit_write(file, buf, count, ppos, &val, ··· 162 162 const char __user *buf, 163 163 size_t count, loff_t *ppos) 164 164 { 165 - unsigned uninitialized_var(val); 165 + unsigned val; 166 166 struct fuse_conn *fc; 167 167 ssize_t ret; 168 168
+1 -1
fs/fuse/cuse.c
··· 270 270 static int cuse_parse_devinfo(char *p, size_t len, struct cuse_devinfo *devinfo) 271 271 { 272 272 char *end = p + len; 273 - char *uninitialized_var(key), *uninitialized_var(val); 273 + char *key, *val; 274 274 int rc; 275 275 276 276 while (true) {
+1 -1
fs/fuse/file.c
··· 2963 2963 { 2964 2964 spin_lock(&fc->lock); 2965 2965 if (RB_EMPTY_NODE(&ff->polled_node)) { 2966 - struct rb_node **link, *uninitialized_var(parent); 2966 + struct rb_node **link, *parent; 2967 2967 2968 2968 link = fuse_find_polled_node(fc, ff->kh, &parent); 2969 2969 BUG_ON(*link);
+1 -1
fs/gfs2/aops.c
··· 335 335 int done = 0; 336 336 struct pagevec pvec; 337 337 int nr_pages; 338 - pgoff_t uninitialized_var(writeback_index); 338 + pgoff_t writeback_index; 339 339 pgoff_t index; 340 340 pgoff_t end; 341 341 pgoff_t done_index;
+1 -1
fs/gfs2/bmap.c
··· 1761 1761 u64 lblock = (offset + (1 << bsize_shift) - 1) >> bsize_shift; 1762 1762 __u16 start_list[GFS2_MAX_META_HEIGHT]; 1763 1763 __u16 __end_list[GFS2_MAX_META_HEIGHT], *end_list = NULL; 1764 - unsigned int start_aligned, uninitialized_var(end_aligned); 1764 + unsigned int start_aligned, end_aligned; 1765 1765 unsigned int strip_h = ip->i_height - 1; 1766 1766 u32 btotal = 0; 1767 1767 int ret, state;
+1 -1
fs/gfs2/lops.c
··· 419 419 struct page *page) 420 420 { 421 421 struct gfs2_sbd *sdp = GFS2_SB(jd->jd_inode); 422 - struct gfs2_log_header_host uninitialized_var(lh); 422 + struct gfs2_log_header_host lh; 423 423 void *kaddr = kmap_atomic(page); 424 424 unsigned int offset; 425 425 bool ret = false;
+1 -1
fs/hfsplus/unicode.c
··· 398 398 astr = str->name; 399 399 len = str->len; 400 400 while (len > 0) { 401 - int uninitialized_var(dsize); 401 + int dsize; 402 402 size = asc2unichar(sb, astr, len, &c); 403 403 astr += size; 404 404 len -= size;
+2 -2
fs/isofs/namei.c
··· 152 152 struct dentry *isofs_lookup(struct inode *dir, struct dentry *dentry, unsigned int flags) 153 153 { 154 154 int found; 155 - unsigned long uninitialized_var(block); 156 - unsigned long uninitialized_var(offset); 155 + unsigned long block; 156 + unsigned long offset; 157 157 struct inode *inode; 158 158 struct page *page; 159 159
+1 -1
fs/jffs2/erase.c
··· 401 401 { 402 402 size_t retlen; 403 403 int ret; 404 - uint32_t uninitialized_var(bad_offset); 404 + uint32_t bad_offset; 405 405 406 406 switch (jffs2_block_check_erase(c, jeb, &bad_offset)) { 407 407 case -EAGAIN: goto refile;
+1 -1
fs/nfsd/nfsctl.c
··· 351 351 static ssize_t write_filehandle(struct file *file, char *buf, size_t size) 352 352 { 353 353 char *dname, *path; 354 - int uninitialized_var(maxsize); 354 + int maxsize; 355 355 char *mesg = buf; 356 356 int len; 357 357 struct auth_domain *dom;
+2 -2
fs/ocfs2/alloc.c
··· 4707 4707 struct ocfs2_alloc_context *meta_ac) 4708 4708 { 4709 4709 int status; 4710 - int uninitialized_var(free_records); 4710 + int free_records; 4711 4711 struct buffer_head *last_eb_bh = NULL; 4712 4712 struct ocfs2_insert_type insert = {0, }; 4713 4713 struct ocfs2_extent_rec rec; ··· 7051 7051 int need_free = 0; 7052 7052 u32 bit_off, num; 7053 7053 handle_t *handle; 7054 - u64 uninitialized_var(block); 7054 + u64 block; 7055 7055 struct ocfs2_inode_info *oi = OCFS2_I(inode); 7056 7056 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 7057 7057 struct ocfs2_dinode *di = (struct ocfs2_dinode *)di_bh->b_data;
+7 -7
fs/ocfs2/dir.c
··· 848 848 u64 *ret_phys_blkno) 849 849 { 850 850 int ret = 0; 851 - unsigned int cend, uninitialized_var(clen); 852 - u32 uninitialized_var(cpos); 853 - u64 uninitialized_var(blkno); 851 + unsigned int cend, clen; 852 + u32 cpos; 853 + u64 blkno; 854 854 u32 name_hash = hinfo->major_hash; 855 855 856 856 ret = ocfs2_dx_dir_lookup_rec(inode, el, name_hash, &cpos, &blkno, ··· 894 894 struct ocfs2_dir_lookup_result *res) 895 895 { 896 896 int ret, i, found; 897 - u64 uninitialized_var(phys); 897 + u64 phys; 898 898 struct buffer_head *dx_leaf_bh = NULL; 899 899 struct ocfs2_dx_leaf *dx_leaf; 900 900 struct ocfs2_dx_entry *dx_entry = NULL; ··· 4393 4393 int ocfs2_dx_dir_truncate(struct inode *dir, struct buffer_head *di_bh) 4394 4394 { 4395 4395 int ret; 4396 - unsigned int uninitialized_var(clen); 4397 - u32 major_hash = UINT_MAX, p_cpos, uninitialized_var(cpos); 4398 - u64 uninitialized_var(blkno); 4396 + unsigned int clen; 4397 + u32 major_hash = UINT_MAX, p_cpos, cpos; 4398 + u64 blkno; 4399 4399 struct ocfs2_super *osb = OCFS2_SB(dir->i_sb); 4400 4400 struct buffer_head *dx_root_bh = NULL; 4401 4401 struct ocfs2_dx_root_block *dx_root;
+2 -2
fs/ocfs2/extent_map.c
··· 403 403 { 404 404 int i, ret, tree_height, len; 405 405 struct ocfs2_dinode *di; 406 - struct ocfs2_extent_block *uninitialized_var(eb); 406 + struct ocfs2_extent_block *eb; 407 407 struct ocfs2_extent_list *el; 408 408 struct ocfs2_extent_rec *rec; 409 409 struct buffer_head *eb_bh = NULL; ··· 599 599 unsigned int *extent_flags) 600 600 { 601 601 int ret; 602 - unsigned int uninitialized_var(hole_len), flags = 0; 602 + unsigned int hole_len, flags = 0; 603 603 struct buffer_head *di_bh = NULL; 604 604 struct ocfs2_extent_rec rec; 605 605
+1 -1
fs/ocfs2/namei.c
··· 2498 2498 struct buffer_head *new_di_bh = NULL; 2499 2499 struct ocfs2_alloc_context *inode_ac = NULL; 2500 2500 struct ocfs2_dir_lookup_result orphan_insert = { NULL, }; 2501 - u64 uninitialized_var(di_blkno), suballoc_loc; 2501 + u64 di_blkno, suballoc_loc; 2502 2502 u16 suballoc_bit; 2503 2503 2504 2504 status = ocfs2_inode_lock(dir, &parent_di_bh, 1);
+1 -1
fs/ocfs2/refcounttree.c
··· 1063 1063 struct buffer_head **ret_bh) 1064 1064 { 1065 1065 int ret = 0, i, found; 1066 - u32 low_cpos, uninitialized_var(cpos_end); 1066 + u32 low_cpos, cpos_end; 1067 1067 struct ocfs2_extent_list *el; 1068 1068 struct ocfs2_extent_rec *rec = NULL; 1069 1069 struct ocfs2_extent_block *eb = NULL;
+1 -1
fs/ocfs2/xattr.c
··· 1211 1211 struct ocfs2_xattr_value_root *xv; 1212 1212 size_t size; 1213 1213 int ret = -ENODATA, name_offset, name_len, i; 1214 - int uninitialized_var(block_off); 1214 + int block_off; 1215 1215 1216 1216 xs->bucket = ocfs2_xattr_bucket_new(inode); 1217 1217 if (!xs->bucket) {
+1 -1
fs/omfs/file.c
··· 220 220 struct buffer_head *bh; 221 221 sector_t next, offset; 222 222 int ret; 223 - u64 uninitialized_var(new_block); 223 + u64 new_block; 224 224 u32 max_extents; 225 225 int extent_count; 226 226 struct omfs_extent *oe;
+1 -1
fs/overlayfs/copy_up.c
··· 787 787 struct path upperpath, datapath; 788 788 int err; 789 789 char *capability = NULL; 790 - ssize_t uninitialized_var(cap_size); 790 + ssize_t cap_size; 791 791 792 792 ovl_path_upper(c->dentry, &upperpath); 793 793 if (WARN_ON(upperpath.dentry == NULL))
+3 -3
fs/ubifs/commit.c
··· 552 552 */ 553 553 int dbg_check_old_index(struct ubifs_info *c, struct ubifs_zbranch *zroot) 554 554 { 555 - int lnum, offs, len, err = 0, uninitialized_var(last_level), child_cnt; 555 + int lnum, offs, len, err = 0, last_level, child_cnt; 556 556 int first = 1, iip; 557 557 struct ubifs_debug_info *d = c->dbg; 558 - union ubifs_key uninitialized_var(lower_key), upper_key, l_key, u_key; 559 - unsigned long long uninitialized_var(last_sqnum); 558 + union ubifs_key lower_key, upper_key, l_key, u_key; 559 + unsigned long long last_sqnum; 560 560 struct ubifs_idx_node *idx; 561 561 struct list_head list; 562 562 struct idx_node *i;
+1 -1
fs/ubifs/dir.c
··· 1260 1260 struct ubifs_budget_req ino_req = { .dirtied_ino = 1, 1261 1261 .dirtied_ino_d = ALIGN(old_inode_ui->data_len, 8) }; 1262 1262 struct timespec64 time; 1263 - unsigned int uninitialized_var(saved_nlink); 1263 + unsigned int saved_nlink; 1264 1264 struct fscrypt_name old_nm, new_nm; 1265 1265 1266 1266 /*
+2 -2
fs/ubifs/file.c
··· 222 222 struct ubifs_info *c = inode->i_sb->s_fs_info; 223 223 pgoff_t index = pos >> PAGE_SHIFT; 224 224 struct ubifs_budget_req req = { .new_page = 1 }; 225 - int uninitialized_var(err), appending = !!(pos + len > inode->i_size); 225 + int err, appending = !!(pos + len > inode->i_size); 226 226 struct page *page; 227 227 228 228 dbg_gen("ino %lu, pos %llu, len %u, i_size %lld", ··· 426 426 struct ubifs_info *c = inode->i_sb->s_fs_info; 427 427 struct ubifs_inode *ui = ubifs_inode(inode); 428 428 pgoff_t index = pos >> PAGE_SHIFT; 429 - int uninitialized_var(err), appending = !!(pos + len > inode->i_size); 429 + int err, appending = !!(pos + len > inode->i_size); 430 430 int skipped_read = 0; 431 431 struct page *page; 432 432
+2 -2
fs/ubifs/journal.c
··· 1222 1222 int aligned_dlen1, aligned_dlen2, plen = UBIFS_INO_NODE_SZ; 1223 1223 int last_reference = !!(new_inode && new_inode->i_nlink == 0); 1224 1224 int move = (old_dir != new_dir); 1225 - struct ubifs_inode *uninitialized_var(new_ui); 1225 + struct ubifs_inode *new_ui; 1226 1226 u8 hash_old_dir[UBIFS_HASH_ARR_SZ]; 1227 1227 u8 hash_new_dir[UBIFS_HASH_ARR_SZ]; 1228 1228 u8 hash_new_inode[UBIFS_HASH_ARR_SZ]; ··· 1507 1507 union ubifs_key key, to_key; 1508 1508 struct ubifs_ino_node *ino; 1509 1509 struct ubifs_trun_node *trun; 1510 - struct ubifs_data_node *uninitialized_var(dn); 1510 + struct ubifs_data_node *dn; 1511 1511 int err, dlen, len, lnum, offs, bit, sz, sync = IS_SYNC(inode); 1512 1512 struct ubifs_inode *ui = ubifs_inode(inode); 1513 1513 ino_t inum = inode->i_ino;
+1 -1
fs/ubifs/lpt.c
··· 275 275 const int k = 32 - nrbits; 276 276 uint8_t *p = *addr; 277 277 int b = *pos; 278 - uint32_t uninitialized_var(val); 278 + uint32_t val; 279 279 const int bytes = (nrbits + b + 7) >> 3; 280 280 281 281 ubifs_assert(c, nrbits > 0);
+3 -3
fs/ubifs/tnc.c
··· 892 892 int adding) 893 893 { 894 894 struct ubifs_znode *o_znode = NULL, *znode = *zn; 895 - int uninitialized_var(o_n), err, cmp, unsure = 0, nn = *n; 895 + int o_n, err, cmp, unsure = 0, nn = *n; 896 896 897 897 cmp = fallible_matches_name(c, &znode->zbranch[nn], nm); 898 898 if (unlikely(cmp < 0)) ··· 1514 1514 */ 1515 1515 int ubifs_tnc_get_bu_keys(struct ubifs_info *c, struct bu_info *bu) 1516 1516 { 1517 - int n, err = 0, lnum = -1, uninitialized_var(offs); 1518 - int uninitialized_var(len); 1517 + int n, err = 0, lnum = -1, offs; 1518 + int len; 1519 1519 unsigned int block = key_block(c, &bu->key); 1520 1520 struct ubifs_znode *znode; 1521 1521
+2 -2
fs/ubifs/tnc_misc.c
··· 126 126 const struct ubifs_znode *znode, 127 127 const union ubifs_key *key, int *n) 128 128 { 129 - int beg = 0, end = znode->child_cnt, uninitialized_var(mid); 130 - int uninitialized_var(cmp); 129 + int beg = 0, end = znode->child_cnt, mid; 130 + int cmp; 131 131 const struct ubifs_zbranch *zbr = &znode->zbranch[0]; 132 132 133 133 ubifs_assert(c, end > beg);
+1 -1
fs/udf/balloc.c
··· 564 564 udf_pblk_t newblock = 0; 565 565 uint32_t adsize; 566 566 uint32_t elen, goal_elen = 0; 567 - struct kernel_lb_addr eloc, uninitialized_var(goal_eloc); 567 + struct kernel_lb_addr eloc, goal_eloc; 568 568 struct extent_position epos, goal_epos; 569 569 int8_t etype; 570 570 struct udf_inode_info *iinfo = UDF_I(table);
+1 -1
fs/xfs/xfs_bmap_util.c
··· 126 126 * pick an extent that will space things out in the rt area. 127 127 */ 128 128 if (ap->eof && ap->offset == 0) { 129 - xfs_rtblock_t uninitialized_var(rtx); /* realtime extent no */ 129 + xfs_rtblock_t rtx; /* realtime extent no */ 130 130 131 131 error = xfs_rtpick_extent(mp, ap->tp, ralen, &rtx); 132 132 if (error)
+1 -1
include/net/flow_offload.h
··· 305 305 struct netlink_ext_ack *extack) 306 306 { 307 307 const struct flow_action_entry *action_entry; 308 - u8 uninitialized_var(last_hw_stats); 308 + u8 last_hw_stats; 309 309 int i; 310 310 311 311 if (flow_offload_has_one_action(action))
+2 -2
kernel/async.c
··· 111 111 struct async_entry *entry = 112 112 container_of(work, struct async_entry, work); 113 113 unsigned long flags; 114 - ktime_t uninitialized_var(calltime), delta, rettime; 114 + ktime_t calltime, delta, rettime; 115 115 116 116 /* 1) run (and print duration) */ 117 117 if (initcall_debug && system_state < SYSTEM_RUNNING) { ··· 287 287 */ 288 288 void async_synchronize_cookie_domain(async_cookie_t cookie, struct async_domain *domain) 289 289 { 290 - ktime_t uninitialized_var(starttime), delta, endtime; 290 + ktime_t starttime, delta, endtime; 291 291 292 292 if (initcall_debug && system_state < SYSTEM_RUNNING) { 293 293 pr_debug("async_waiting @ %i\n", task_pid_nr(current));
+1 -1
kernel/audit.c
··· 1800 1800 { 1801 1801 struct audit_buffer *ab; 1802 1802 struct timespec64 t; 1803 - unsigned int uninitialized_var(serial); 1803 + unsigned int serial; 1804 1804 1805 1805 if (audit_initialized != AUDIT_INITIALIZED) 1806 1806 return NULL;
+1 -1
kernel/debug/kdb/kdb_io.c
··· 591 591 int this_cpu, old_cpu; 592 592 char *cp, *cp2, *cphold = NULL, replaced_byte = ' '; 593 593 char *moreprompt = "more> "; 594 - unsigned long uninitialized_var(flags); 594 + unsigned long flags; 595 595 596 596 /* Serialize kdb_printf if multiple cpus try to write at once. 597 597 * But if any cpu goes recursive in kdb, just print the output,
+1 -1
kernel/dma/debug.c
··· 882 882 static int dma_debug_device_change(struct notifier_block *nb, unsigned long action, void *data) 883 883 { 884 884 struct device *dev = data; 885 - struct dma_debug_entry *uninitialized_var(entry); 885 + struct dma_debug_entry *entry; 886 886 int count; 887 887 888 888 if (dma_debug_disabled())
+1 -1
kernel/events/core.c
··· 11483 11483 struct perf_event *group_leader = NULL, *output_event = NULL; 11484 11484 struct perf_event *event, *sibling; 11485 11485 struct perf_event_attr attr; 11486 - struct perf_event_context *ctx, *uninitialized_var(gctx); 11486 + struct perf_event_context *ctx, *gctx; 11487 11487 struct file *event_file = NULL; 11488 11488 struct fd group = {NULL, 0}; 11489 11489 struct task_struct *task = NULL;
+1 -1
kernel/events/uprobes.c
··· 2189 2189 { 2190 2190 struct uprobe *uprobe; 2191 2191 unsigned long bp_vaddr; 2192 - int uninitialized_var(is_swbp); 2192 + int is_swbp; 2193 2193 2194 2194 bp_vaddr = uprobe_get_swbp_addr(regs); 2195 2195 if (bp_vaddr == get_trampoline_vaddr())
+1 -1
kernel/exit.c
··· 93 93 struct signal_struct *sig = tsk->signal; 94 94 bool group_dead = thread_group_leader(tsk); 95 95 struct sighand_struct *sighand; 96 - struct tty_struct *uninitialized_var(tty); 96 + struct tty_struct *tty; 97 97 u64 utime, stime; 98 98 99 99 sighand = rcu_dereference_check(tsk->sighand,
+7 -7
kernel/futex.c
··· 1326 1326 static int lock_pi_update_atomic(u32 __user *uaddr, u32 uval, u32 newval) 1327 1327 { 1328 1328 int err; 1329 - u32 uninitialized_var(curval); 1329 + u32 curval; 1330 1330 1331 1331 if (unlikely(should_fail_futex(true))) 1332 1332 return -EFAULT; ··· 1496 1496 */ 1497 1497 static int wake_futex_pi(u32 __user *uaddr, u32 uval, struct futex_pi_state *pi_state) 1498 1498 { 1499 - u32 uninitialized_var(curval), newval; 1499 + u32 curval, newval; 1500 1500 struct task_struct *new_owner; 1501 1501 bool postunlock = false; 1502 1502 DEFINE_WAKE_Q(wake_q); ··· 2370 2370 struct task_struct *argowner) 2371 2371 { 2372 2372 struct futex_pi_state *pi_state = q->pi_state; 2373 - u32 uval, uninitialized_var(curval), newval; 2373 + u32 uval, curval, newval; 2374 2374 struct task_struct *oldowner, *newowner; 2375 2375 u32 newtid; 2376 2376 int ret, err = 0; ··· 2996 2996 */ 2997 2997 static int futex_unlock_pi(u32 __user *uaddr, unsigned int flags) 2998 2998 { 2999 - u32 uninitialized_var(curval), uval, vpid = task_pid_vnr(current); 2999 + u32 curval, uval, vpid = task_pid_vnr(current); 3000 3000 union futex_key key = FUTEX_KEY_INIT; 3001 3001 struct futex_hash_bucket *hb; 3002 3002 struct futex_q *top_waiter; ··· 3479 3479 static int handle_futex_death(u32 __user *uaddr, struct task_struct *curr, 3480 3480 bool pi, bool pending_op) 3481 3481 { 3482 - u32 uval, uninitialized_var(nval), mval; 3482 + u32 uval, nval, mval; 3483 3483 int err; 3484 3484 3485 3485 /* Futex address must be 32bit aligned */ ··· 3609 3609 struct robust_list_head __user *head = curr->robust_list; 3610 3610 struct robust_list __user *entry, *next_entry, *pending; 3611 3611 unsigned int limit = ROBUST_LIST_LIMIT, pi, pip; 3612 - unsigned int uninitialized_var(next_pi); 3612 + unsigned int next_pi; 3613 3613 unsigned long futex_offset; 3614 3614 int rc; 3615 3615 ··· 3909 3909 struct compat_robust_list_head __user *head = curr->compat_robust_list; 3910 3910 struct robust_list __user *entry, *next_entry, *pending; 3911 3911 unsigned int limit = ROBUST_LIST_LIMIT, pi, pip; 3912 - unsigned int uninitialized_var(next_pi); 3912 + unsigned int next_pi; 3913 3913 compat_uptr_t uentry, next_uentry, upending; 3914 3914 compat_long_t futex_offset; 3915 3915 int rc;
+8 -8
kernel/locking/lockdep.c
··· 1723 1723 static unsigned long __lockdep_count_forward_deps(struct lock_list *this) 1724 1724 { 1725 1725 unsigned long count = 0; 1726 - struct lock_list *uninitialized_var(target_entry); 1726 + struct lock_list *target_entry; 1727 1727 1728 1728 __bfs_forwards(this, (void *)&count, noop_count, &target_entry); 1729 1729 ··· 1749 1749 static unsigned long __lockdep_count_backward_deps(struct lock_list *this) 1750 1750 { 1751 1751 unsigned long count = 0; 1752 - struct lock_list *uninitialized_var(target_entry); 1752 + struct lock_list *target_entry; 1753 1753 1754 1754 __bfs_backwards(this, (void *)&count, noop_count, &target_entry); 1755 1755 ··· 1804 1804 struct lock_trace **const trace) 1805 1805 { 1806 1806 int ret; 1807 - struct lock_list *uninitialized_var(target_entry); 1807 + struct lock_list *target_entry; 1808 1808 struct lock_list src_entry = { 1809 1809 .class = hlock_class(src), 1810 1810 .parent = NULL, ··· 1842 1842 check_redundant(struct held_lock *src, struct held_lock *target) 1843 1843 { 1844 1844 int ret; 1845 - struct lock_list *uninitialized_var(target_entry); 1845 + struct lock_list *target_entry; 1846 1846 struct lock_list src_entry = { 1847 1847 .class = hlock_class(src), 1848 1848 .parent = NULL, ··· 2244 2244 { 2245 2245 unsigned long usage_mask = 0, forward_mask, backward_mask; 2246 2246 enum lock_usage_bit forward_bit = 0, backward_bit = 0; 2247 - struct lock_list *uninitialized_var(target_entry1); 2248 - struct lock_list *uninitialized_var(target_entry); 2247 + struct lock_list *target_entry1; 2248 + struct lock_list *target_entry; 2249 2249 struct lock_list this, that; 2250 2250 int ret; 2251 2251 ··· 3438 3438 { 3439 3439 int ret; 3440 3440 struct lock_list root; 3441 - struct lock_list *uninitialized_var(target_entry); 3441 + struct lock_list *target_entry; 3442 3442 3443 3443 root.parent = NULL; 3444 3444 root.class = hlock_class(this); ··· 3465 3465 { 3466 3466 int ret; 3467 3467 struct lock_list root; 3468 - struct lock_list *uninitialized_var(target_entry); 3468 + struct lock_list *target_entry; 3469 3469 3470 3470 root.parent = NULL; 3471 3471 root.class = hlock_class(this);
+1 -1
kernel/trace/ring_buffer.c
··· 577 577 */ 578 578 int ring_buffer_wait(struct trace_buffer *buffer, int cpu, int full) 579 579 { 580 - struct ring_buffer_per_cpu *uninitialized_var(cpu_buffer); 580 + struct ring_buffer_per_cpu *cpu_buffer; 581 581 DEFINE_WAIT(wait); 582 582 struct rb_irq_work *work; 583 583 int ret = 0;
+1 -1
lib/radix-tree.c
··· 1029 1029 { 1030 1030 struct radix_tree_node *node, *parent; 1031 1031 unsigned long maxindex; 1032 - int uninitialized_var(offset); 1032 + int offset; 1033 1033 1034 1034 radix_tree_load_root(root, &node, &maxindex); 1035 1035 if (index > maxindex)
+1 -1
lib/test_lockup.c
··· 168 168 169 169 static void test_lock(bool master, bool verbose) 170 170 { 171 - u64 uninitialized_var(wait_start); 171 + u64 wait_start; 172 172 173 173 if (measure_lock_wait) 174 174 wait_start = local_clock();
+1 -1
mm/frontswap.c
··· 446 446 void frontswap_shrink(unsigned long target_pages) 447 447 { 448 448 unsigned long pages_to_unuse = 0; 449 - int uninitialized_var(type), ret; 449 + int type, ret; 450 450 451 451 /* 452 452 * we don't want to hold swap_lock while doing a very
+1 -1
mm/ksm.c
··· 2387 2387 static void ksm_do_scan(unsigned int scan_npages) 2388 2388 { 2389 2389 struct rmap_item *rmap_item; 2390 - struct page *uninitialized_var(page); 2390 + struct page *page; 2391 2391 2392 2392 while (scan_npages-- && likely(!freezing(current))) { 2393 2393 cond_resched();
+1 -1
mm/memcontrol.c
··· 1004 1004 struct mem_cgroup *prev, 1005 1005 struct mem_cgroup_reclaim_cookie *reclaim) 1006 1006 { 1007 - struct mem_cgroup_reclaim_iter *uninitialized_var(iter); 1007 + struct mem_cgroup_reclaim_iter *iter; 1008 1008 struct cgroup_subsys_state *css = NULL; 1009 1009 struct mem_cgroup *memcg = NULL; 1010 1010 struct mem_cgroup *pos = NULL;
+1 -1
mm/memory.c
··· 2205 2205 { 2206 2206 pte_t *pte; 2207 2207 int err = 0; 2208 - spinlock_t *uninitialized_var(ptl); 2208 + spinlock_t *ptl; 2209 2209 2210 2210 if (create) { 2211 2211 pte = (mm == &init_mm) ?
+2 -2
mm/mempolicy.c
··· 1234 1234 static struct page *new_page(struct page *page, unsigned long start) 1235 1235 { 1236 1236 struct vm_area_struct *vma; 1237 - unsigned long uninitialized_var(address); 1237 + unsigned long address; 1238 1238 1239 1239 vma = find_vma(current->mm, start); 1240 1240 while (vma) { ··· 1629 1629 unsigned long flags) 1630 1630 { 1631 1631 int err; 1632 - int uninitialized_var(pval); 1632 + int pval; 1633 1633 nodemask_t nodes; 1634 1634 1635 1635 addr = untagged_addr(addr);
+1 -1
mm/page_alloc.c
··· 961 961 int migratetype, bool report) 962 962 { 963 963 struct capture_control *capc = task_capc(zone); 964 - unsigned long uninitialized_var(buddy_pfn); 964 + unsigned long buddy_pfn; 965 965 unsigned long combined_pfn; 966 966 unsigned int max_order; 967 967 struct page *buddy;
+1 -1
mm/percpu.c
··· 2513 2513 const size_t static_size = __per_cpu_end - __per_cpu_start; 2514 2514 int nr_groups = 1, nr_units = 0; 2515 2515 size_t size_sum, min_unit_size, alloc_size; 2516 - int upa, max_upa, uninitialized_var(best_upa); /* units_per_alloc */ 2516 + int upa, max_upa, best_upa; /* units_per_alloc */ 2517 2517 int last_allocs, group, unit; 2518 2518 unsigned int cpu, tcpu; 2519 2519 struct pcpu_alloc_info *ai;
+2 -2
mm/slub.c
··· 1218 1218 struct kmem_cache_node *n = get_node(s, page_to_nid(page)); 1219 1219 void *object = head; 1220 1220 int cnt = 0; 1221 - unsigned long uninitialized_var(flags); 1221 + unsigned long flags; 1222 1222 int ret = 0; 1223 1223 1224 1224 spin_lock_irqsave(&n->list_lock, flags); ··· 2901 2901 struct page new; 2902 2902 unsigned long counters; 2903 2903 struct kmem_cache_node *n = NULL; 2904 - unsigned long uninitialized_var(flags); 2904 + unsigned long flags; 2905 2905 2906 2906 stat(s, FREE_SLOWPATH); 2907 2907
+2 -2
mm/swap.c
··· 830 830 LIST_HEAD(pages_to_free); 831 831 struct pglist_data *locked_pgdat = NULL; 832 832 struct lruvec *lruvec; 833 - unsigned long uninitialized_var(flags); 834 - unsigned int uninitialized_var(lock_batch); 833 + unsigned long flags; 834 + unsigned int lock_batch; 835 835 836 836 for (i = 0; i < nr; i++) { 837 837 struct page *page = pages[i];
+1 -1
net/dccp/options.c
··· 56 56 (dh->dccph_doff * 4); 57 57 struct dccp_options_received *opt_recv = &dp->dccps_options_received; 58 58 unsigned char opt, len; 59 - unsigned char *uninitialized_var(value); 59 + unsigned char *value; 60 60 u32 elapsed_time; 61 61 __be32 opt_val; 62 62 int rc;
+3 -3
net/ipv4/netfilter/nf_socket_ipv4.c
··· 84 84 struct sock *nf_sk_lookup_slow_v4(struct net *net, const struct sk_buff *skb, 85 85 const struct net_device *indev) 86 86 { 87 - __be32 uninitialized_var(daddr), uninitialized_var(saddr); 88 - __be16 uninitialized_var(dport), uninitialized_var(sport); 87 + __be32 daddr, saddr; 88 + __be16 dport, sport; 89 89 const struct iphdr *iph = ip_hdr(skb); 90 90 struct sk_buff *data_skb = NULL; 91 - u8 uninitialized_var(protocol); 91 + u8 protocol; 92 92 #if IS_ENABLED(CONFIG_NF_CONNTRACK) 93 93 enum ip_conntrack_info ctinfo; 94 94 struct nf_conn const *ct;
+1 -1
net/ipv6/ip6_flowlabel.c
··· 535 535 536 536 int ipv6_flowlabel_opt(struct sock *sk, char __user *optval, int optlen) 537 537 { 538 - int uninitialized_var(err); 538 + int err; 539 539 struct net *net = sock_net(sk); 540 540 struct ipv6_pinfo *np = inet6_sk(sk); 541 541 struct in6_flowlabel_req freq;
+1 -1
net/ipv6/netfilter/nf_socket_ipv6.c
··· 97 97 struct sock *nf_sk_lookup_slow_v6(struct net *net, const struct sk_buff *skb, 98 98 const struct net_device *indev) 99 99 { 100 - __be16 uninitialized_var(dport), uninitialized_var(sport); 100 + __be16 dport, sport; 101 101 const struct in6_addr *daddr = NULL, *saddr = NULL; 102 102 struct ipv6hdr *iph = ipv6_hdr(skb); 103 103 struct sk_buff *data_skb = NULL;
+1 -1
net/netfilter/nf_conntrack_ftp.c
··· 382 382 int ret; 383 383 u32 seq; 384 384 int dir = CTINFO2DIR(ctinfo); 385 - unsigned int uninitialized_var(matchlen), uninitialized_var(matchoff); 385 + unsigned int matchlen, matchoff; 386 386 struct nf_ct_ftp_master *ct_ftp_info = nfct_help_data(ct); 387 387 struct nf_conntrack_expect *exp; 388 388 union nf_inet_addr *daddr;
+1 -1
net/sched/cls_flow.c
··· 225 225 226 226 static u32 flow_get_vlan_tag(const struct sk_buff *skb) 227 227 { 228 - u16 uninitialized_var(tag); 228 + u16 tag; 229 229 230 230 if (vlan_get_tag(skb, &tag) < 0) 231 231 return 0;
+1 -1
net/sched/sch_cake.c
··· 1692 1692 { 1693 1693 struct cake_sched_data *q = qdisc_priv(sch); 1694 1694 int len = qdisc_pkt_len(skb); 1695 - int uninitialized_var(ret); 1695 + int ret; 1696 1696 struct sk_buff *ack = NULL; 1697 1697 ktime_t now = ktime_get(); 1698 1698 struct cake_tin_data *b;
+1 -1
net/sched/sch_cbq.c
··· 360 360 struct sk_buff **to_free) 361 361 { 362 362 struct cbq_sched_data *q = qdisc_priv(sch); 363 - int uninitialized_var(ret); 363 + int ret; 364 364 struct cbq_class *cl = cbq_classify(skb, sch, &ret); 365 365 366 366 #ifdef CONFIG_NET_CLS_ACT
+1 -1
net/sched/sch_fq_codel.c
··· 187 187 struct fq_codel_sched_data *q = qdisc_priv(sch); 188 188 unsigned int idx, prev_backlog, prev_qlen; 189 189 struct fq_codel_flow *flow; 190 - int uninitialized_var(ret); 190 + int ret; 191 191 unsigned int pkt_len; 192 192 bool memory_limited; 193 193
+1 -1
net/sched/sch_fq_pie.c
··· 130 130 { 131 131 struct fq_pie_sched_data *q = qdisc_priv(sch); 132 132 struct fq_pie_flow *sel_flow; 133 - int uninitialized_var(ret); 133 + int ret; 134 134 u8 memory_limited = false; 135 135 u8 enqueue = false; 136 136 u32 pkt_len;
+1 -1
net/sched/sch_hfsc.c
··· 1533 1533 { 1534 1534 unsigned int len = qdisc_pkt_len(skb); 1535 1535 struct hfsc_class *cl; 1536 - int uninitialized_var(err); 1536 + int err; 1537 1537 bool first; 1538 1538 1539 1539 cl = hfsc_classify(skb, sch, &err);
+1 -1
net/sched/sch_htb.c
··· 579 579 static int htb_enqueue(struct sk_buff *skb, struct Qdisc *sch, 580 580 struct sk_buff **to_free) 581 581 { 582 - int uninitialized_var(ret); 582 + int ret; 583 583 unsigned int len = qdisc_pkt_len(skb); 584 584 struct htb_sched *q = qdisc_priv(sch); 585 585 struct htb_class *cl = htb_classify(skb, sch, &ret);
+1 -1
net/sched/sch_sfq.c
··· 349 349 unsigned int hash, dropped; 350 350 sfq_index x, qlen; 351 351 struct sfq_slot *slot; 352 - int uninitialized_var(ret); 352 + int ret; 353 353 struct sk_buff *head; 354 354 int delta; 355 355
+2 -2
net/sunrpc/svcsock.c
··· 563 563 .msg_control = cmh, 564 564 .msg_controllen = sizeof(buffer), 565 565 }; 566 - unsigned int uninitialized_var(sent); 566 + unsigned int sent; 567 567 int err; 568 568 569 569 svc_udp_release_rqst(rqstp); ··· 1080 1080 struct msghdr msg = { 1081 1081 .msg_flags = 0, 1082 1082 }; 1083 - unsigned int uninitialized_var(sent); 1083 + unsigned int sent; 1084 1084 int err; 1085 1085 1086 1086 svc_tcp_release_rqst(rqstp);
+5 -5
net/sunrpc/xprtsock.c
··· 496 496 int flags, struct rpc_rqst *req) 497 497 { 498 498 struct xdr_buf *buf = &req->rq_private_buf; 499 - size_t want, uninitialized_var(read); 500 - ssize_t uninitialized_var(ret); 499 + size_t want, read; 500 + ssize_t ret; 501 501 502 502 xs_read_header(transport, buf); 503 503 ··· 844 844 struct msghdr msg = { 845 845 .msg_flags = XS_SENDMSG_FLAGS, 846 846 }; 847 - unsigned int uninitialized_var(sent); 847 + unsigned int sent; 848 848 int status; 849 849 850 850 /* Close the stream if the previous transmission was incomplete */ ··· 915 915 .msg_namelen = xprt->addrlen, 916 916 .msg_flags = XS_SENDMSG_FLAGS, 917 917 }; 918 - unsigned int uninitialized_var(sent); 918 + unsigned int sent; 919 919 int status; 920 920 921 921 xs_pktdump("packet data:", ··· 999 999 .msg_flags = XS_SENDMSG_FLAGS, 1000 1000 }; 1001 1001 bool vm_wait = false; 1002 - unsigned int uninitialized_var(sent); 1002 + unsigned int sent; 1003 1003 int status; 1004 1004 1005 1005 /* Close the stream if the previous transmission was incomplete */
+1 -1
net/tls/tls_sw.c
··· 670 670 struct tls_prot_info *prot = &tls_ctx->prot_info; 671 671 struct tls_sw_context_tx *ctx = tls_sw_ctx_tx(tls_ctx); 672 672 struct tls_rec *rec = ctx->open_rec, *tmp = NULL; 673 - u32 i, split_point, uninitialized_var(orig_end); 673 + u32 i, split_point, orig_end; 674 674 struct sk_msg *msg_pl, *msg_en; 675 675 struct aead_request *req; 676 676 bool split;
+1 -1
sound/core/control_compat.c
··· 223 223 { 224 224 struct snd_ctl_elem_value32 __user *data32 = userdata; 225 225 int i, type, size; 226 - int uninitialized_var(count); 226 + int count; 227 227 unsigned int indirect; 228 228 229 229 if (copy_from_user(&data->id, &data32->id, sizeof(data->id)))
+1 -1
sound/isa/sb/sb16_csp.c
··· 102 102 int snd_sb_csp_new(struct snd_sb *chip, int device, struct snd_hwdep ** rhwdep) 103 103 { 104 104 struct snd_sb_csp *p; 105 - int uninitialized_var(version); 105 + int version; 106 106 int err; 107 107 struct snd_hwdep *hw; 108 108
+1 -1
sound/usb/endpoint.c
··· 335 335 while (test_bit(EP_FLAG_RUNNING, &ep->flags)) { 336 336 337 337 unsigned long flags; 338 - struct snd_usb_packet_info *uninitialized_var(packet); 338 + struct snd_usb_packet_info *packet; 339 339 struct snd_urb_ctx *ctx = NULL; 340 340 int err, i; 341 341