From bac73556c98fb94d5c790c211b7e535f878ca85a Mon Sep 17 00:00:00 2001 From: "google-labs-jules[bot]" <161369871+google-labs-jules[bot]@users.noreply.github.com> Date: Wed, 18 Feb 2026 05:03:24 +0000 Subject: Fix unbounded memory usage in crawler (DoS) Co-authored-by: adammathes <868470+adammathes@users.noreply.github.com> --- models/item/item_bench_test.go | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'models/item') diff --git a/models/item/item_bench_test.go b/models/item/item_bench_test.go index ba674c7..b904c32 100644 --- a/models/item/item_bench_test.go +++ b/models/item/item_bench_test.go @@ -200,10 +200,10 @@ func BenchmarkFilter_LargeDataset(b *testing.B) { if i > 0 { sb.WriteString(",") } - fmt.Fprintf(&sb, + sb.WriteString(fmt.Sprintf( "('Item %d', 'https://example.com/large/%d', '

Description %d

', datetime('now'), %d, 0, 0)", i, i, i, feedID, - ) + )) } _, err := models.DB.Exec( "INSERT INTO item(title, url, description, publish_date, feed_id, read_state, starred) VALUES " + sb.String(), -- cgit v1.2.3