{"api_version":"1","generated_at":"2026-05-12T22:16:59+00:00","cve":"CVE-2026-23294","urls":{"html":"https://cve.report/CVE-2026-23294","api":"https://cve.report/api/cve/CVE-2026-23294.json","docs":"https://cve.report/api","cve_org":"https://www.cve.org/CVERecord?id=CVE-2026-23294","nvd":"https://nvd.nist.gov/vuln/detail/CVE-2026-23294"},"summary":{"title":"bpf: Fix race in devmap on PREEMPT_RT","description":"In the Linux kernel, the following vulnerability has been resolved:\n\nbpf: Fix race in devmap on PREEMPT_RT\n\nOn PREEMPT_RT kernels, the per-CPU xdp_dev_bulk_queue (bq) can be\naccessed concurrently by multiple preemptible tasks on the same CPU.\n\nThe original code assumes bq_enqueue() and __dev_flush() run atomically\nwith respect to each other on the same CPU, relying on\nlocal_bh_disable() to prevent preemption. However, on PREEMPT_RT,\nlocal_bh_disable() only calls migrate_disable() (when\nPREEMPT_RT_NEEDS_BH_LOCK is not set) and does not disable\npreemption, which allows CFS scheduling to preempt a task during\nbq_xmit_all(), enabling another task on the same CPU to enter\nbq_enqueue() and operate on the same per-CPU bq concurrently.\n\nThis leads to several races:\n\n1. Double-free / use-after-free on bq->q[]: bq_xmit_all() snapshots\n   cnt = bq->count, then iterates bq->q[0..cnt-1] to transmit frames.\n   If preempted after the snapshot, a second task can call bq_enqueue()\n   -> bq_xmit_all() on the same bq, transmitting (and freeing) the\n   same frames. When the first task resumes, it operates on stale\n   pointers in bq->q[], causing use-after-free.\n\n2. bq->count and bq->q[] corruption: concurrent bq_enqueue() modifying\n   bq->count and bq->q[] while bq_xmit_all() is reading them.\n\n3. dev_rx/xdp_prog teardown race: __dev_flush() clears bq->dev_rx and\n   bq->xdp_prog after bq_xmit_all(). If preempted between\n   bq_xmit_all() return and bq->dev_rx = NULL, a preempting\n   bq_enqueue() sees dev_rx still set (non-NULL), skips adding bq to\n   the flush_list, and enqueues a frame. When __dev_flush() resumes,\n   it clears dev_rx and removes bq from the flush_list, orphaning the\n   newly enqueued frame.\n\n4. __list_del_clearprev() on flush_node: similar to the cpumap race,\n   both tasks can call __list_del_clearprev() on the same flush_node,\n   the second dereferences the prev pointer already set to NULL.\n\nThe race between task A (__dev_flush -> bq_xmit_all) and task B\n(bq_enqueue -> bq_xmit_all) on the same CPU:\n\n  Task A (xdp_do_flush)          Task B (ndo_xdp_xmit redirect)\n  ----------------------         --------------------------------\n  __dev_flush(flush_list)\n    bq_xmit_all(bq)\n      cnt = bq->count  /* e.g. 16 */\n      /* start iterating bq->q[] */\n    <-- CFS preempts Task A -->\n                                   bq_enqueue(dev, xdpf)\n                                     bq->count == DEV_MAP_BULK_SIZE\n                                     bq_xmit_all(bq, 0)\n                                       cnt = bq->count  /* same 16! */\n                                       ndo_xdp_xmit(bq->q[])\n                                       /* frames freed by driver */\n                                       bq->count = 0\n    <-- Task A resumes -->\n      ndo_xdp_xmit(bq->q[])\n      /* use-after-free: frames already freed! */\n\nFix this by adding a local_lock_t to xdp_dev_bulk_queue and acquiring\nit in bq_enqueue() and __dev_flush(). These paths already run under\nlocal_bh_disable(), so use local_lock_nested_bh() which on non-RT is\na pure annotation with no overhead, and on PREEMPT_RT provides a\nper-CPU sleeping lock that serializes access to the bq.","state":"PUBLISHED","assigner":"Linux","published_at":"2026-03-25 11:16:24","updated_at":"2026-04-02 15:16:30"},"problem_types":[],"metrics":[{"version":"3.1","source":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","type":"Secondary","score":"7","severity":"HIGH","vector":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","data":{"version":"3.1","vectorString":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","baseScore":7,"baseSeverity":"HIGH","attackVector":"LOCAL","attackComplexity":"HIGH","privilegesRequired":"LOW","userInteraction":"NONE","scope":"UNCHANGED","confidentialityImpact":"HIGH","integrityImpact":"HIGH","availabilityImpact":"HIGH"}},{"version":"3.1","source":"CNA","type":"DECLARED","score":"7","severity":"HIGH","vector":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","data":{"baseScore":7,"baseSeverity":"HIGH","vectorString":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","version":"3.1"}}],"references":[{"url":"https://git.kernel.org/stable/c/6c10b019785dc282c5f45d21e4a3f468b8fd6476","name":"https://git.kernel.org/stable/c/6c10b019785dc282c5f45d21e4a3f468b8fd6476","refsource":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","tags":[],"title":"","mime":"","httpstatus":"","archivestatus":"0"},{"url":"https://git.kernel.org/stable/c/1872e75375c40add4a35990de3be77b5741c252c","name":"https://git.kernel.org/stable/c/1872e75375c40add4a35990de3be77b5741c252c","refsource":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","tags":[],"title":"","mime":"","httpstatus":"","archivestatus":"0"},{"url":"https://git.kernel.org/stable/c/ab1a56c9d99189aa5c6e03940d06e40ba6a28240","name":"https://git.kernel.org/stable/c/ab1a56c9d99189aa5c6e03940d06e40ba6a28240","refsource":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","tags":[],"title":"","mime":"","httpstatus":"","archivestatus":"0"},{"url":"https://www.cve.org/CVERecord?id=CVE-2026-23294","name":"CVE Program record","refsource":"CVE.ORG","tags":["canonical"]},{"url":"https://nvd.nist.gov/vuln/detail/CVE-2026-23294","name":"NVD vulnerability detail","refsource":"NVD","tags":["canonical","analysis"]}],"affected":[{"source":"CNA","vendor":"Linux","product":"Linux","version":"affected 3253cb49cbad4772389d6ef55be75db1f97da910 6c10b019785dc282c5f45d21e4a3f468b8fd6476 git","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"affected 3253cb49cbad4772389d6ef55be75db1f97da910 ab1a56c9d99189aa5c6e03940d06e40ba6a28240 git","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"affected 3253cb49cbad4772389d6ef55be75db1f97da910 1872e75375c40add4a35990de3be77b5741c252c git","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"affected 6.18","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"unaffected 6.18 semver","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"unaffected 6.18.17 6.18.* semver","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"unaffected 6.19.7 6.19.* semver","platforms":[]},{"source":"CNA","vendor":"Linux","product":"Linux","version":"unaffected 7.0-rc2 * original_commit_for_fix","platforms":[]}],"timeline":[],"solutions":[],"workarounds":[],"exploits":[],"credits":[],"nvd_cpes":[],"vendor_comments":[],"enrichments":{"kev":null,"epss":{"cve_year":"2026","cve_id":"23294","cve":"CVE-2026-23294","epss":"0.000130000","percentile":"0.021630000","score_date":"2026-04-07","updated_at":"2026-04-08 00:03:40"},"legacy_qids":[]},"source_records":{"cve_program":{"containers":{"cna":{"affected":[{"defaultStatus":"unaffected","product":"Linux","programFiles":["kernel/bpf/devmap.c"],"repo":"https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git","vendor":"Linux","versions":[{"lessThan":"6c10b019785dc282c5f45d21e4a3f468b8fd6476","status":"affected","version":"3253cb49cbad4772389d6ef55be75db1f97da910","versionType":"git"},{"lessThan":"ab1a56c9d99189aa5c6e03940d06e40ba6a28240","status":"affected","version":"3253cb49cbad4772389d6ef55be75db1f97da910","versionType":"git"},{"lessThan":"1872e75375c40add4a35990de3be77b5741c252c","status":"affected","version":"3253cb49cbad4772389d6ef55be75db1f97da910","versionType":"git"}]},{"defaultStatus":"affected","product":"Linux","programFiles":["kernel/bpf/devmap.c"],"repo":"https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git","vendor":"Linux","versions":[{"status":"affected","version":"6.18"},{"lessThan":"6.18","status":"unaffected","version":"0","versionType":"semver"},{"lessThanOrEqual":"6.18.*","status":"unaffected","version":"6.18.17","versionType":"semver"},{"lessThanOrEqual":"6.19.*","status":"unaffected","version":"6.19.7","versionType":"semver"},{"lessThanOrEqual":"*","status":"unaffected","version":"7.0-rc2","versionType":"original_commit_for_fix"}]}],"cpeApplicability":[{"nodes":[{"cpeMatch":[{"criteria":"cpe:2.3:o:linux:linux_kernel:*:*:*:*:*:*:*:*","versionEndExcluding":"6.18.17","versionStartIncluding":"6.18","vulnerable":true},{"criteria":"cpe:2.3:o:linux:linux_kernel:*:*:*:*:*:*:*:*","versionEndExcluding":"6.19.7","versionStartIncluding":"6.18","vulnerable":true},{"criteria":"cpe:2.3:o:linux:linux_kernel:*:*:*:*:*:*:*:*","versionEndExcluding":"7.0-rc2","versionStartIncluding":"6.18","vulnerable":true}],"negate":false,"operator":"OR"}]}],"descriptions":[{"lang":"en","value":"In the Linux kernel, the following vulnerability has been resolved:\n\nbpf: Fix race in devmap on PREEMPT_RT\n\nOn PREEMPT_RT kernels, the per-CPU xdp_dev_bulk_queue (bq) can be\naccessed concurrently by multiple preemptible tasks on the same CPU.\n\nThe original code assumes bq_enqueue() and __dev_flush() run atomically\nwith respect to each other on the same CPU, relying on\nlocal_bh_disable() to prevent preemption. However, on PREEMPT_RT,\nlocal_bh_disable() only calls migrate_disable() (when\nPREEMPT_RT_NEEDS_BH_LOCK is not set) and does not disable\npreemption, which allows CFS scheduling to preempt a task during\nbq_xmit_all(), enabling another task on the same CPU to enter\nbq_enqueue() and operate on the same per-CPU bq concurrently.\n\nThis leads to several races:\n\n1. Double-free / use-after-free on bq->q[]: bq_xmit_all() snapshots\n   cnt = bq->count, then iterates bq->q[0..cnt-1] to transmit frames.\n   If preempted after the snapshot, a second task can call bq_enqueue()\n   -> bq_xmit_all() on the same bq, transmitting (and freeing) the\n   same frames. When the first task resumes, it operates on stale\n   pointers in bq->q[], causing use-after-free.\n\n2. bq->count and bq->q[] corruption: concurrent bq_enqueue() modifying\n   bq->count and bq->q[] while bq_xmit_all() is reading them.\n\n3. dev_rx/xdp_prog teardown race: __dev_flush() clears bq->dev_rx and\n   bq->xdp_prog after bq_xmit_all(). If preempted between\n   bq_xmit_all() return and bq->dev_rx = NULL, a preempting\n   bq_enqueue() sees dev_rx still set (non-NULL), skips adding bq to\n   the flush_list, and enqueues a frame. When __dev_flush() resumes,\n   it clears dev_rx and removes bq from the flush_list, orphaning the\n   newly enqueued frame.\n\n4. __list_del_clearprev() on flush_node: similar to the cpumap race,\n   both tasks can call __list_del_clearprev() on the same flush_node,\n   the second dereferences the prev pointer already set to NULL.\n\nThe race between task A (__dev_flush -> bq_xmit_all) and task B\n(bq_enqueue -> bq_xmit_all) on the same CPU:\n\n  Task A (xdp_do_flush)          Task B (ndo_xdp_xmit redirect)\n  ----------------------         --------------------------------\n  __dev_flush(flush_list)\n    bq_xmit_all(bq)\n      cnt = bq->count  /* e.g. 16 */\n      /* start iterating bq->q[] */\n    <-- CFS preempts Task A -->\n                                   bq_enqueue(dev, xdpf)\n                                     bq->count == DEV_MAP_BULK_SIZE\n                                     bq_xmit_all(bq, 0)\n                                       cnt = bq->count  /* same 16! */\n                                       ndo_xdp_xmit(bq->q[])\n                                       /* frames freed by driver */\n                                       bq->count = 0\n    <-- Task A resumes -->\n      ndo_xdp_xmit(bq->q[])\n      /* use-after-free: frames already freed! */\n\nFix this by adding a local_lock_t to xdp_dev_bulk_queue and acquiring\nit in bq_enqueue() and __dev_flush(). These paths already run under\nlocal_bh_disable(), so use local_lock_nested_bh() which on non-RT is\na pure annotation with no overhead, and on PREEMPT_RT provides a\nper-CPU sleeping lock that serializes access to the bq."}],"metrics":[{"cvssV3_1":{"baseScore":7,"baseSeverity":"HIGH","vectorString":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","version":"3.1"}}],"providerMetadata":{"dateUpdated":"2026-04-02T14:44:14.096Z","orgId":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","shortName":"Linux"},"references":[{"url":"https://git.kernel.org/stable/c/6c10b019785dc282c5f45d21e4a3f468b8fd6476"},{"url":"https://git.kernel.org/stable/c/ab1a56c9d99189aa5c6e03940d06e40ba6a28240"},{"url":"https://git.kernel.org/stable/c/1872e75375c40add4a35990de3be77b5741c252c"}],"title":"bpf: Fix race in devmap on PREEMPT_RT","x_generator":{"engine":"bippy-1.2.0"}}},"cveMetadata":{"assignerOrgId":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","assignerShortName":"Linux","cveId":"CVE-2026-23294","datePublished":"2026-03-25T10:26:51.946Z","dateReserved":"2026-01-13T15:37:45.993Z","dateUpdated":"2026-04-02T14:44:14.096Z","state":"PUBLISHED"},"dataType":"CVE_RECORD","dataVersion":"5.2"},"nvd":{"publishedDate":"2026-03-25 11:16:24","lastModifiedDate":"2026-04-02 15:16:30","problem_types":[],"metrics":{"cvssMetricV31":[{"source":"416baaa9-dc9f-4396-8d5f-8c081fb06d67","type":"Secondary","cvssData":{"version":"3.1","vectorString":"CVSS:3.1/AV:L/AC:H/PR:L/UI:N/S:U/C:H/I:H/A:H","baseScore":7,"baseSeverity":"HIGH","attackVector":"LOCAL","attackComplexity":"HIGH","privilegesRequired":"LOW","userInteraction":"NONE","scope":"UNCHANGED","confidentialityImpact":"HIGH","integrityImpact":"HIGH","availabilityImpact":"HIGH"},"exploitabilityScore":1,"impactScore":5.9}]},"configurations":[]},"legacy_mitre":{"record":{"CveYear":"2026","CveId":"23294","Ordinal":"1","Title":"bpf: Fix race in devmap on PREEMPT_RT","CVE":"CVE-2026-23294","Year":"2026"},"notes":[{"CveYear":"2026","CveId":"23294","Ordinal":"1","NoteData":"In the Linux kernel, the following vulnerability has been resolved:\n\nbpf: Fix race in devmap on PREEMPT_RT\n\nOn PREEMPT_RT kernels, the per-CPU xdp_dev_bulk_queue (bq) can be\naccessed concurrently by multiple preemptible tasks on the same CPU.\n\nThe original code assumes bq_enqueue() and __dev_flush() run atomically\nwith respect to each other on the same CPU, relying on\nlocal_bh_disable() to prevent preemption. However, on PREEMPT_RT,\nlocal_bh_disable() only calls migrate_disable() (when\nPREEMPT_RT_NEEDS_BH_LOCK is not set) and does not disable\npreemption, which allows CFS scheduling to preempt a task during\nbq_xmit_all(), enabling another task on the same CPU to enter\nbq_enqueue() and operate on the same per-CPU bq concurrently.\n\nThis leads to several races:\n\n1. Double-free / use-after-free on bq->q[]: bq_xmit_all() snapshots\n   cnt = bq->count, then iterates bq->q[0..cnt-1] to transmit frames.\n   If preempted after the snapshot, a second task can call bq_enqueue()\n   -> bq_xmit_all() on the same bq, transmitting (and freeing) the\n   same frames. When the first task resumes, it operates on stale\n   pointers in bq->q[], causing use-after-free.\n\n2. bq->count and bq->q[] corruption: concurrent bq_enqueue() modifying\n   bq->count and bq->q[] while bq_xmit_all() is reading them.\n\n3. dev_rx/xdp_prog teardown race: __dev_flush() clears bq->dev_rx and\n   bq->xdp_prog after bq_xmit_all(). If preempted between\n   bq_xmit_all() return and bq->dev_rx = NULL, a preempting\n   bq_enqueue() sees dev_rx still set (non-NULL), skips adding bq to\n   the flush_list, and enqueues a frame. When __dev_flush() resumes,\n   it clears dev_rx and removes bq from the flush_list, orphaning the\n   newly enqueued frame.\n\n4. __list_del_clearprev() on flush_node: similar to the cpumap race,\n   both tasks can call __list_del_clearprev() on the same flush_node,\n   the second dereferences the prev pointer already set to NULL.\n\nThe race between task A (__dev_flush -> bq_xmit_all) and task B\n(bq_enqueue -> bq_xmit_all) on the same CPU:\n\n  Task A (xdp_do_flush)          Task B (ndo_xdp_xmit redirect)\n  ----------------------         --------------------------------\n  __dev_flush(flush_list)\n    bq_xmit_all(bq)\n      cnt = bq->count  /* e.g. 16 */\n      /* start iterating bq->q[] */\n    <-- CFS preempts Task A -->\n                                   bq_enqueue(dev, xdpf)\n                                     bq->count == DEV_MAP_BULK_SIZE\n                                     bq_xmit_all(bq, 0)\n                                       cnt = bq->count  /* same 16! */\n                                       ndo_xdp_xmit(bq->q[])\n                                       /* frames freed by driver */\n                                       bq->count = 0\n    <-- Task A resumes -->\n      ndo_xdp_xmit(bq->q[])\n      /* use-after-free: frames already freed! */\n\nFix this by adding a local_lock_t to xdp_dev_bulk_queue and acquiring\nit in bq_enqueue() and __dev_flush(). These paths already run under\nlocal_bh_disable(), so use local_lock_nested_bh() which on non-RT is\na pure annotation with no overhead, and on PREEMPT_RT provides a\nper-CPU sleeping lock that serializes access to the bq.","Type":"Description","Title":"bpf: Fix race in devmap on PREEMPT_RT"}]}}}