官术网_书友最值得收藏!

Storing data using AWS S3

There are many cases where we just want to save content that we scrape into a local copy for archive purposes, backup, or later bulk analysis. We also might want to save media from those sites for later use. I've built scrapers for advertisement compliance companies, where we would track and download advertisement based media on web sites to ensure proper usage, and also to store for later analysis, compliance and transcoding.

The storage required for these types of systems can be immense, but with the advent of cloud storage services such as AWS S3 (Simple Storage Service), this becomes much easier and more cost effective than managing a large SAN (Storage Area Network) in your own IT department. Plus, S3 can also automatically move data from hot to cold storage, and then to long-term storage, such as a glacier, which can save you much more money.

We won't get into all of those details, but simply look at storing our planets.html file into an S3 bucket. Once you can do this, you can save any content you want to year hearts desire.

主站蜘蛛池模板: 收藏| 江津市| 松溪县| 盐山县| 鸡西市| 宝清县| 乌恰县| 龙山县| 汕尾市| 兴安县| 渝中区| 雷山县| 平南县| 奉化市| 广河县| 延庆县| 吉隆县| 宜章县| 河西区| 吴江市| 孟津县| 前郭尔| 东明县| 新竹市| 肥东县| 钟山县| 玛纳斯县| 兴和县| 额敏县| 繁昌县| 葫芦岛市| 沾益县| 崇礼县| 明溪县| 黄浦区| 包头市| 成安县| 游戏| 庆城县| 高台县| 临高县|