Fixed placeholder text

This commit is contained in:
CocoByte 2024-02-28 20:13:30 -07:00
parent b9bebef6d5
commit 54b27ef492
No known key found for this signature in database
GPG key ID: BBFAA2526384C97F

View file

@ -44,7 +44,7 @@ NOTE: While we considered implementing caching in a sandbox (See footnote) in or
Therefore, implementing caching using Whitenoise is not currently worth it for the following reasons;
- Minimal gains: We would only be caching static files (total load time gain estimated to be….)
- Risks: Incurs risk of unforeseen loading issues (we cant entirely rule out that we wont run into issues like we did in our xx-xx-xx incident). Although we dont think static files should pose a problem, due diligence would call us to monitor for any unforeseen issues that might arise, which adds cost to this project that doesnt seem proportional to the gains.
- Risks: Incurs risk of unforeseen loading issues (we cant entirely rule out that we wont run into issues like we did in our November launch incident). Although we dont think static files should pose a problem, due diligence would call us to monitor for any unforeseen issues that might arise, which adds cost to this project that doesnt seem proportional to the gains.
- Maintenance: We would have to provide custom settings in cloudfront (coordinated through Cameron) for any sandboxes and other environments where caching is enabled. If we move down the route of utilizing CDN, it would be good for every environment to have this service enabled so our dev environments reflect stable settings. This could possibly introduce some overhead and maintenance issues. (Although further investigation might reveal these to be negligible.)
Overall, it is recommended that we SHELVE this caching endeavor for a future scenario where we have exhausted other (likely more lucrative) options for performance improvements. If we then still need to make improvements to our load times, perhaps we can revisit this and examine caching not only static files, but other resources as well (with caution).