diff --git a/oauthproxy.go b/oauthproxy.go index 588d827a..e2b89ec7 100644 --- a/oauthproxy.go +++ b/oauthproxy.go @@ -577,13 +577,7 @@ func (p *OAuthProxy) serveHTTP(rw http.ResponseWriter, req *http.Request) { // RobotsTxt disallows scraping pages from the OAuthProxy func (p *OAuthProxy) RobotsTxt(rw http.ResponseWriter, req *http.Request) { - _, err := fmt.Fprintf(rw, "User-agent: *\nDisallow: /") - if err != nil { - logger.Printf("Error writing robots.txt: %v", err) - p.ErrorPage(rw, req, http.StatusInternalServerError, err.Error()) - return - } - rw.WriteHeader(http.StatusOK) + p.pageWriter.WriteRobotsTxt(rw, req) } // ErrorPage writes an error response diff --git a/oauthproxy_test.go b/oauthproxy_test.go index a2805c34..42b9b042 100644 --- a/oauthproxy_test.go +++ b/oauthproxy_test.go @@ -57,7 +57,7 @@ func TestRobotsTxt(t *testing.T) { req, _ := http.NewRequest("GET", "/robots.txt", nil) proxy.ServeHTTP(rw, req) assert.Equal(t, 200, rw.Code) - assert.Equal(t, "User-agent: *\nDisallow: /", rw.Body.String()) + assert.Equal(t, "User-agent: *\nDisallow: /\n", rw.Body.String()) } func TestIsValidRedirect(t *testing.T) {