| 2026-03-20 22:29 UTC |
rcu: Use an intermediate irq_work to start process_srcu() |
2 |
boqun@kernel.org |
skipped
|
| 2026-03-20 18:14 UTC |
rcu: Use an intermediate irq_work to start process_srcu() |
1 |
boqun@kernel.org |
skipped
|
| 2026-03-18 09:16 UTC |
tools build: Append -fzero-init-padding-bits=all option |
5 |
leo.yan@arm.com |
finished
in 1h0m0s
|
| 2026-03-18 08:53 UTC |
tools/sched_ext: Update stale scx_ops_error() comment in fcg_cgroup_move() |
3 |
ke.zhao.kernel@gmail.com |
finished
in 1h18m0s
|
| 2026-03-18 06:14 UTC |
tools/sched_ext: Update stale scx_ops_error comment in fcg_cgroup_move |
2 |
ke.zhao.kernel@gmail.com |
finished
in 1h12m0s
|
| 2026-03-18 06:02 UTC |
tools/sched_ext: Update stale scx_ops_error comment in fcg_cgroup_move |
1 |
ke.zhao.kernel@gmail.com |
finished
in 1h11m0s
|
| 2026-03-16 22:27 UTC |
bpf: Always defer local storage free |
1 |
arighi@nvidia.com |
finished
in 3h32m0s
|
| 2026-03-11 08:29 UTC |
tools build: Append -fzero-init-padding-bits=all option |
4 |
leo.yan@arm.com |
finished
in 1h1m0s
|
| 2026-03-10 11:48 UTC |
treewide: Use IS_ERR_OR_NULL over manual NULL check - refactor |
1 |
phahn-oss@avm.de |
finished
in 4h4m0s
|
| 2026-03-09 02:28 UTC |
sched_ext cleanup and toolchain compatibility fixes |
2 |
zhaomzhao@126.com |
skipped
|
| 2026-03-08 16:46 UTC |
tools build: Append -fzero-init-padding-bits=all option |
3 |
leo.yan@arm.com |
finished
in 52m0s
|
| 2026-03-07 00:54 UTC |
bpf: refine u32/s32 bounds when ranges cross min/max boundary |
3 |
eddyz87@gmail.com |
finished
in 4h12m0s
|
| 2026-03-06 07:31 UTC |
sched_ext cleanup and toolchain compatibility fixes |
1 |
zhaomzhao@126.com |
finished
in 1h17m0s
|
| 2026-03-05 19:48 UTC |
bpf: refine u32/s32 bounds when ranges cross min/max boundary |
2 |
eddyz87@gmail.com |
finished
in 4h0m0s
|
| 2026-03-05 11:09 UTC |
bpf: refine u32/s32 bounds when ranges cross min/max boundary |
1 |
eddyz87@gmail.com |
finished
in 4h33m0s
|
| 2026-03-04 05:37 UTC |
sched_ext: Use WRITE_ONCE() for the write side of dsq->seq update |
1 |
soolaugust@gmail.com |
finished
in 1h16m0s
|
| 2026-03-03 07:23 UTC |
tools/sched_ext: fix compilation issues |
1 |
zhaomzhao@126.com |
finished
in 1h7m0s
|
| 2026-02-20 19:07 UTC |
bpf: Prevent invalid u32 bounds in __reg32_deduce_bounds() |
1 |
arighi@nvidia.com |
finished
in 4h10m0s
|
| 2026-02-13 18:21 UTC |
selftests/sched_ext: Fix rt_stall flaky failure |
1 |
ihor.solodrai@linux.dev |
finished
in 59m0s
|
| 2026-02-09 08:55 UTC |
tools/sched_ext: Use u64 for kind field to fix ARM64 atomic ops |
1 |
xuewen.yan@unisoc.com |
finished
in 51m0s
|
| 2025-12-31 17:36 UTC |
bpf: use const pointer for struct_ops cfi_stubs |
1 |
csander@purestorage.com |
finished
in 1h5m0s
|
| 2025-12-19 01:59 UTC |
resolve_btfids: Support for BTF modifications |
6 |
ihor.solodrai@linux.dev |
finished
in 56m0s
|
| 2025-12-19 00:31 UTC |
resolve_btfids: Support for BTF modifications |
5 |
ihor.solodrai@linux.dev |
finished
in 56m0s
|
| 2025-12-18 00:33 UTC |
resolve_btfids: Support for BTF modifications |
4 |
ihor.solodrai@linux.dev |
finished
in 49m0s
|
| 2025-09-24 08:14 UTC |
bpf: Mark kfuncs as __noclone |
2 |
arighi@nvidia.com |
finished
in 42m0s
|
| 2025-09-23 09:16 UTC |
sched_ext: Verify RCU protection in scx_bpf_cpu_curr() |
1 |
arighi@nvidia.com |
finished
in 51m0s
|
| 2025-09-17 03:27 UTC |
Update KF_RCU_PROTECTED |
3 |
memxor@gmail.com |
finished
in 3h45m0s
|
| 2025-09-17 03:20 UTC |
Update KF_RCU_PROTECTED |
2 |
memxor@gmail.com |
finished
in 3h46m0s
|
| 2025-09-15 02:47 UTC |
Update KF_RCU_PROTECTED, add KF_RET_RCU |
1 |
memxor@gmail.com |
finished
in 3h39m0s
|
| 2025-08-22 14:05 UTC |
bpf: Mark kfuncs as __noclone |
1 |
arighi@nvidia.com |
finished
in 37m0s
|