Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

503 Errors from CPANTesters website #10

Open
preaction opened this issue May 3, 2016 · 3 comments
Open

503 Errors from CPANTesters website #10

preaction opened this issue May 3, 2016 · 3 comments
Labels

Comments

@preaction
Copy link
Member

When we added Fastly to get caching and improve performance, we started getting intermittent 503 backend read error messages. We need to make sure the Fastly caching is working, figure out why the 503 errors are happening, and try to reduce them.

@preaction
Copy link
Member Author

preaction commented May 5, 2016

On the Fastly website (http://fastly.com), I've updated the timeouts for "First bytes" to 60,000 ms (60s). This should drastically reduce the amount of 503s we get. We should check back with people in a few days to see if they've gotten any more 503s.

@ghost
Copy link

ghost commented Jun 24, 2016

@preaction This is still an issue. Just got the following:

Error 503 backend read error

backend read error
Guru Mediation:

Details: cache-dfw1823-DFW 1466788517 1422049808

@preaction
Copy link
Member Author

preaction commented Sep 23, 2016

@wchristian just reported this again as well. This time it seems that the view-report.cgi was being hammered (as usual). @mst has volunteered to try something interesting, but otherwise has also suggested Plack::Handler::WrapCGI with its execute flag to do a pre-forking CGI script that saves us from having to load a bunch of modules every time. If we restrict it to just the view-report.cgi for now, we can see how this helps us before deciding if we need to roll it out to more of the CGI scripts.

During the present errors I also noticed that we were spending 40-90% CPU time in iowait. I checked iostat and it seems like there's a lot of writing to disk going on. I probably need to track down where the writes are coming from and see if any can be reduced. We've got a lot of free memory, which concerns me that it isn't being used for disk i/o caching. I know we've got a lot of various log files being written, which it might be possible to reduce using syslog. But I also enabled the data release (#9) which will certainly increase the amount of disk we're writing at various times. It might be necessary to move the website to another machine, separate from the database and backend processes...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant