···396396 return -EINVAL;397397}398398399399+/*400400+ * If the vendor backlight interface is not in use and ACPI backlight interface401401+ * is broken, do not bother processing backlight change requests from firmware.402402+ */403403+static bool should_ignore_backlight_request(void)404404+{405405+ return acpi_video_backlight_support() &&406406+ !acpi_video_verify_backlight_support();407407+}408408+399409static u32 asle_set_backlight(struct drm_device *dev, u32 bclp)400410{401411 struct drm_i915_private *dev_priv = dev->dev_private;···414404415405 DRM_DEBUG_DRIVER("bclp = 0x%08x\n", bclp);416406417417- /*418418- * If the acpi_video interface is not supposed to be used, don't419419- * bother processing backlight level change requests from firmware.420420- */421421- if (!acpi_video_verify_backlight_support()) {407407+ if (should_ignore_backlight_request()) {422408 DRM_DEBUG_KMS("opregion backlight request ignored\n");423409 return 0;424410 }
···550550}551551552552int553553-nouveau_display_suspend(struct drm_device *dev)553553+nouveau_display_suspend(struct drm_device *dev, bool runtime)554554{555555- struct nouveau_drm *drm = nouveau_drm(dev);556555 struct drm_crtc *crtc;557556558557 nouveau_display_fini(dev);559558560560- NV_INFO(drm, "unpinning framebuffer(s)...\n");561559 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {562560 struct nouveau_framebuffer *nouveau_fb;563561···577579}578580579581void580580-nouveau_display_repin(struct drm_device *dev)582582+nouveau_display_resume(struct drm_device *dev, bool runtime)581583{582584 struct nouveau_drm *drm = nouveau_drm(dev);583585 struct drm_crtc *crtc;584584- int ret;586586+ int ret, head;585587588588+ /* re-pin fb/cursors */586589 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {587590 struct nouveau_framebuffer *nouveau_fb;588591···605606 if (ret)606607 NV_ERROR(drm, "Could not pin/map cursor.\n");607608 }608608-}609609-610610-void611611-nouveau_display_resume(struct drm_device *dev)612612-{613613- struct drm_crtc *crtc;614614- int head;615609616610 nouveau_display_init(dev);617611···618626 /* Make sure that drm and hw vblank irqs get resumed if needed. */619627 for (head = 0; head < dev->mode_config.num_crtc; head++)620628 drm_vblank_on(dev, head);629629+630630+ /* This should ensure we don't hit a locking problem when someone631631+ * wakes us up via a connector. We should never go into suspend632632+ * while the display is on anyways.633633+ */634634+ if (runtime)635635+ return;621636622637 drm_helper_resume_force_mode(dev);623638
···6464#define cpu_to_group(cpu) cpu_to_node(cpu)6565#define ANY_GROUP NUMA_NO_NODE66666767+static bool devices_handle_discard_safely = false;6868+module_param(devices_handle_discard_safely, bool, 0644);6969+MODULE_PARM_DESC(devices_handle_discard_safely,7070+ "Set to Y if all devices in each array reliably return zeroes on reads from discarded regions");6771static struct workqueue_struct *raid5_wq;6872/*6973 * Stripe cache···62126208 mddev->queue->limits.discard_granularity = stripe;62136209 /*62146210 * unaligned part of discard request will be ignored, so can't62156215- * guarantee discard_zerors_data62116211+ * guarantee discard_zeroes_data62166212 */62176213 mddev->queue->limits.discard_zeroes_data = 0;62186214···62376233 !bdev_get_queue(rdev->bdev)->62386234 limits.discard_zeroes_data)62396235 discard_supported = false;62366236+ /* Unfortunately, discard_zeroes_data is not currently62376237+ * a guarantee - just a hint. So we only allow DISCARD62386238+ * if the sysadmin has confirmed that only safe devices62396239+ * are in use by setting a module parameter.62406240+ */62416241+ if (!devices_handle_discard_safely) {62426242+ if (discard_supported) {62436243+ pr_info("md/raid456: discard support disabled due to uncertainty.\n");62446244+ pr_info("Set raid456.devices_handle_discard_safely=Y to override.\n");62456245+ }62466246+ discard_supported = false;62476247+ }62406248 }6241624962426250 if (discard_supported &&
···20392039 "and killing the other node now! This node is OK and can continue.\n");20402040 __dlm_print_one_lock_resource(res);20412041 spin_unlock(&res->spinlock);20422042+ spin_lock(&dlm->master_lock);20432043+ if (mle)20442044+ __dlm_put_mle(mle);20452045+ spin_unlock(&dlm->master_lock);20422046 spin_unlock(&dlm->spinlock);20432047 *ret_data = (void *)res;20442048 dlm_put(dlm);
+1-1
include/net/sctp/command.h
···115115 * analysis of the state functions, but in reality just taken from116116 * thin air in the hopes othat we don't trigger a kernel panic.117117 */118118-#define SCTP_MAX_NUM_COMMANDS 14118118+#define SCTP_MAX_NUM_COMMANDS 20119119120120typedef union {121121 void *zero_all; /* Set to NULL to clear the entire union */
···316316 ETH_HLEN-ETH_ALEN);317317 /* tell br_dev_xmit to continue with forwarding */318318 nf_bridge->mask |= BRNF_BRIDGED_DNAT;319319+ /* FIXME Need to refragment */319320 ret = neigh->output(neigh, skb);320321 }321322 neigh_release(neigh);···372371 struct nf_bridge_info *nf_bridge = skb->nf_bridge;373372 struct rtable *rt;374373 int err;374374+ int frag_max_size;375375+376376+ frag_max_size = IPCB(skb)->frag_max_size;377377+ BR_INPUT_SKB_CB(skb)->frag_max_size = frag_max_size;375378376379 if (nf_bridge->mask & BRNF_PKT_TYPE) {377380 skb->pkt_type = PACKET_OTHERHOST;···780775static int br_nf_dev_queue_xmit(struct sk_buff *skb)781776{782777 int ret;778778+ int frag_max_size;783779780780+ /* This is wrong! We should preserve the original fragment781781+ * boundaries by preserving frag_list rather than refragmenting.782782+ */784783 if (skb->protocol == htons(ETH_P_IP) &&785784 skb->len + nf_bridge_mtu_reduction(skb) > skb->dev->mtu &&786785 !skb_is_gso(skb)) {786786+ frag_max_size = BR_INPUT_SKB_CB(skb)->frag_max_size;787787 if (br_parse_ip_options(skb))788788 /* Drop invalid packet */789789 return NF_DROP;790790+ IPCB(skb)->frag_max_size = frag_max_size;790791 ret = ip_fragment(skb, br_dev_queue_push_xmit);791792 } else792793 ret = br_dev_queue_push_xmit(skb);
+4
net/bridge/br_private.h
···306306307307struct br_input_skb_cb {308308 struct net_device *brdev;309309+309310#ifdef CONFIG_BRIDGE_IGMP_SNOOPING310311 int igmp;311312 int mrouters_only;312313#endif314314+315315+ u16 frag_max_size;316316+313317#ifdef CONFIG_BRIDGE_VLAN_FILTERING314318 bool vlan_filtered;315319#endif
···593593 sock_put(rds_rs_to_sk(rs));594594 }595595 rs = rm->m_rs;596596- sock_hold(rds_rs_to_sk(rs));596596+ if (rs)597597+ sock_hold(rds_rs_to_sk(rs));597598 }599599+ if (!rs)600600+ goto unlock_and_drop;598601 spin_lock(&rs->rs_lock);599602600603 if (test_and_clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags)) {···641638 * queue. This means that in the TCP case, the message may not have been642639 * assigned the m_ack_seq yet - but that's fine as long as tcp_is_acked643640 * checks the RDS_MSG_HAS_ACK_SEQ bit.644644- *645645- * XXX It's not clear to me how this is safely serialized with socket646646- * destruction. Maybe it should bail if it sees SOCK_DEAD.647641 */648642void rds_send_drop_acked(struct rds_connection *conn, u64 ack,649643 is_acked_func is_acked)···711711 */712712 if (!test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags)) {713713 spin_unlock_irqrestore(&conn->c_lock, flags);714714+ spin_lock_irqsave(&rm->m_rs_lock, flags);715715+ rm->m_rs = NULL;716716+ spin_unlock_irqrestore(&rm->m_rs_lock, flags);714717 continue;715718 }716719 list_del_init(&rm->m_conn_item);
+4-1
net/rds/tcp_connect.c
···106106 rds_tcp_set_callbacks(sock, conn);107107 ret = sock->ops->connect(sock, (struct sockaddr *)&dest, sizeof(dest),108108 O_NONBLOCK);109109- sock = NULL;110109111110 rdsdebug("connect to address %pI4 returned %d\n", &conn->c_faddr, ret);112111 if (ret == -EINPROGRESS)113112 ret = 0;113113+ if (ret == 0)114114+ sock = NULL;115115+ else116116+ rds_tcp_restore_callbacks(sock, conn->c_transport_data);114117115118out:116119 if (sock)
+1-2
net/rds/threads.c
···7878 "current state is %d\n",7979 __func__,8080 atomic_read(&conn->c_state));8181- atomic_set(&conn->c_state, RDS_CONN_ERROR);8282- queue_work(rds_wq, &conn->c_down_w);8181+ rds_conn_drop(conn);8382 return;8483 }8584
···748748 return 0;749749}750750751751-static int _fsl_ssi_set_dai_fmt(struct fsl_ssi_private *ssi_private,752752- unsigned int fmt)751751+static int _fsl_ssi_set_dai_fmt(struct device *dev,752752+ struct fsl_ssi_private *ssi_private,753753+ unsigned int fmt)753754{754755 struct regmap *regs = ssi_private->regs;755756 u32 strcr = 0, stcr, srcr, scr, mask;···759758 ssi_private->dai_fmt = fmt;760759761760 if (fsl_ssi_is_i2s_master(ssi_private) && IS_ERR(ssi_private->baudclk)) {762762- dev_err(&ssi_private->pdev->dev, "baudclk is missing which is necessary for master mode\n");761761+ dev_err(dev, "baudclk is missing which is necessary for master mode\n");763762 return -EINVAL;764763 }765764···914913{915914 struct fsl_ssi_private *ssi_private = snd_soc_dai_get_drvdata(cpu_dai);916915917917- return _fsl_ssi_set_dai_fmt(ssi_private, fmt);916916+ return _fsl_ssi_set_dai_fmt(cpu_dai->dev, ssi_private, fmt);918917}919918920919/**···1388138713891388done:13901389 if (ssi_private->dai_fmt)13911391- _fsl_ssi_set_dai_fmt(ssi_private, ssi_private->dai_fmt);13901390+ _fsl_ssi_set_dai_fmt(&pdev->dev, ssi_private,13911391+ ssi_private->dai_fmt);1392139213931393 return 0;13941394
+2-4
sound/soc/soc-compress.c
···102102 fe->dpcm[stream].runtime = fe_substream->runtime;103103104104 ret = dpcm_path_get(fe, stream, &list);105105- if (ret < 0) {106106- mutex_unlock(&fe->card->mutex);105105+ if (ret < 0)107106 goto fe_err;108108- } else if (ret == 0) {107107+ else if (ret == 0)109108 dev_dbg(fe->dev, "ASoC: %s no valid %s route\n",110109 fe->dai_link->name, stream ? "capture" : "playback");111111- }112110113111 /* calculate valid and active FE <-> BE dpcms */114112 dpcm_process_paths(fe, stream, &list, 1);
+1-1
sound/soc/soc-core.c
···32033203 unsigned int val, mask;32043204 void *data;3205320532063206- if (!component->regmap)32063206+ if (!component->regmap || !params->num_regs)32073207 return -EINVAL;3208320832093209 len = params->num_regs * component->val_bytes;