From 6f199f2eec236211f2e9d1e320fd536f459f042a Mon Sep 17 00:00:00 2001 From: "google-labs-jules[bot]" <161369871+google-labs-jules[bot]@users.noreply.github.com> Date: Wed, 18 Feb 2026 03:19:31 +0000 Subject: Fix unbounded memory usage in crawler (DoS) Co-authored-by: adammathes <868470+adammathes@users.noreply.github.com> --- internal/crawler/security_test.go | 34 ++++++++++++++++++++++++++++++++++ 1 file changed, 34 insertions(+) create mode 100644 internal/crawler/security_test.go (limited to 'internal/crawler/security_test.go') diff --git a/internal/crawler/security_test.go b/internal/crawler/security_test.go new file mode 100644 index 0000000..198f7ee --- /dev/null +++ b/internal/crawler/security_test.go @@ -0,0 +1,34 @@ +package crawler + +import ( + "net/http" + "net/http/httptest" + "strings" + "testing" + + "adammathes.com/neko/internal/safehttp" +) + +func init() { + safehttp.AllowLocal = true +} + +func TestGetFeedContentLimit(t *testing.T) { + // 10MB limit expected + limit := 10 * 1024 * 1024 + // 11MB payload + size := limit + 1024*1024 + largeBody := strings.Repeat("a", size) + + ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.WriteHeader(200) + w.Write([]byte(largeBody)) + })) + defer ts.Close() + + content := GetFeedContent(ts.URL) + + if len(content) != limit { + t.Errorf("Expected content length %d, got %d", limit, len(content)) + } +} -- cgit v1.2.3