What are some strategies for scaling sites & infrastructure so global response times are relatively close to US response times?
18th October 2013
My answer to What are some strategies for scaling sites & infrastructure so global response times are relatively close to US response times? on Quora
You need to run your application in multiple data centers around the world, partitioned such that an incoming HTTP request can be completely serviced by a single data center. Then you use global DNS load balancing to direct users to the data center that is closest to them.
Building an application like this is extremely difficult due to the need to synchronise data between your data centers, and to avoid inconsistencies if those data centers lose connectivity between them. Most startups avoid doing this until they have large engineering teams and can afford to hire people who have done this before.
If you want to speed up read-only traffic this kind of thing is easier to achieve—you can work with a CDN/edge caching company such as http://fastly.com/, Akamai or Amazon CloudFront who run their own servers around the world and can cache your content for you. This can dramatically speed up your site for international visitors.
More recent articles
- Weeknotes: Llama 3, AI for Data Journalism, llm-evals and datasette-secrets - 23rd April 2024
- Options for accessing Llama 3 from the terminal using LLM - 22nd April 2024
- AI for Data Journalism: demonstrating what we can do with this stuff right now - 17th April 2024
- Three major LLM releases in 24 hours (plus weeknotes) - 10th April 2024
- Building files-to-prompt entirely using Claude 3 Opus - 8th April 2024
- Running OCR against PDFs and images directly in your browser - 30th March 2024
- llm cmd undo last git commit - a new plugin for LLM - 26th March 2024
- Building and testing C extensions for SQLite with ChatGPT Code Interpreter - 23rd March 2024
- Claude and ChatGPT for ad-hoc sidequests - 22nd March 2024
- Weeknotes: the aftermath of NICAR - 16th March 2024