high_scalability high_scalability-2009 high_scalability-2009-525 knowledge-graph by maker-knowledge-mining
Source: html
Introduction: Update: HostedFTP.com - Amazon S3 Performance Report . How fast is S3? Based on their own study HostedFTP.com has found: 10 to 12 MB/second when storing and receiving files and 140 ms per file stored as a fixed overhead cost. Update: A Quantitative Comparison of Rackspace and Amazon Cloud Storage Solutions . S3 isn't the only cloud storage service out there. Mosso is saying they can save you so money while offering support. There are number of scenarios in their paper, but For 5TB of cloud storage Mosso will save you 17% over S3 without support and 42% with support. For their CDN on a Global test Mosso says the average response time is 333ms for CloudFront vs. 107ms for Cloud Files which means globally, Cloud Files is 3.1 times or 211% faster than CloudFront. Amazon S3 is storage for the Internet. It is designed to make web-scale computing easier for developers. This service allows you to link directly to files at a cost of 15 cents per GB of storage, and 20 cents per GB
sentIndex sentText sentNum sentScore
1 com has found: 10 to 12 MB/second when storing and receiving files and 140 ms per file stored as a fixed overhead cost. [sent-5, score-1.081]
2 S3 isn't the only cloud storage service out there. [sent-7, score-0.458]
3 Mosso is saying they can save you so money while offering support. [sent-8, score-0.423]
4 There are number of scenarios in their paper, but For 5TB of cloud storage Mosso will save you 17% over S3 without support and 42% with support. [sent-9, score-0.73]
5 For their CDN on a Global test Mosso says the average response time is 333ms for CloudFront vs. [sent-10, score-0.285]
6 107ms for Cloud Files which means globally, Cloud Files is 3. [sent-11, score-0.053]
7 It is designed to make web-scale computing easier for developers. [sent-14, score-0.194]
8 This service allows you to link directly to files at a cost of 15 cents per GB of storage, and 20 cents per GB transfer. [sent-15, score-1.464]
wordName wordTfidf (topN-words)
[('mosso', 0.581), ('files', 0.3), ('cents', 0.286), ('gb', 0.201), ('storage', 0.19), ('cloud', 0.189), ('quantitative', 0.167), ('cloudfront', 0.16), ('save', 0.147), ('receiving', 0.146), ('transfer', 0.121), ('ms', 0.119), ('globally', 0.118), ('per', 0.118), ('rackspace', 0.118), ('scenarios', 0.115), ('comparison', 0.111), ('study', 0.11), ('offering', 0.103), ('amazon', 0.1), ('saying', 0.098), ('fixed', 0.097), ('link', 0.097), ('overhead', 0.094), ('cdn', 0.09), ('storing', 0.082), ('average', 0.08), ('service', 0.079), ('easier', 0.078), ('says', 0.078), ('money', 0.075), ('paper', 0.072), ('directly', 0.071), ('stored', 0.068), ('global', 0.066), ('response', 0.064), ('test', 0.063), ('found', 0.062), ('designed', 0.062), ('update', 0.061), ('allows', 0.059), ('file', 0.057), ('faster', 0.056), ('computing', 0.054), ('means', 0.053), ('times', 0.052), ('cost', 0.05), ('support', 0.047), ('number', 0.042), ('fast', 0.042)]
simIndex simValue blogId blogTitle
same-blog 1 1.0000001 525 high scalability-2009-03-05-Product: Amazon Simple Storage Service
Introduction: Update: HostedFTP.com - Amazon S3 Performance Report . How fast is S3? Based on their own study HostedFTP.com has found: 10 to 12 MB/second when storing and receiving files and 140 ms per file stored as a fixed overhead cost. Update: A Quantitative Comparison of Rackspace and Amazon Cloud Storage Solutions . S3 isn't the only cloud storage service out there. Mosso is saying they can save you so money while offering support. There are number of scenarios in their paper, but For 5TB of cloud storage Mosso will save you 17% over S3 without support and 42% with support. For their CDN on a Global test Mosso says the average response time is 333ms for CloudFront vs. 107ms for Cloud Files which means globally, Cloud Files is 3.1 times or 211% faster than CloudFront. Amazon S3 is storage for the Internet. It is designed to make web-scale computing easier for developers. This service allows you to link directly to files at a cost of 15 cents per GB of storage, and 20 cents per GB
2 0.21673256 195 high scalability-2007-12-28-Amazon's EC2: Pay as You Grow Could Cut Your Costs in Half
Introduction: Update 2: Summize Computes Computing Resources for a Startup . Lots of nice graphs showing Amazon is hard to beat for small machines and become less cost efficient for well used larger machines. Long term storage costs may eat your saving away. And out of cloud bandwidth costs are high. Update: via ProductionScale , a nice Digital Web article on how to setup S3 to store media files and how Blue Origin was able to handle 3.5 million requests and 758 GBs in bandwidth in a single day for very little $$$. Also a Right Scale article on Network performance within Amazon EC2 and to Amazon S3 . 75MB/s between EC2 instances, 10.2MB/s between EC2 and S3 for download, 6.9MB/s upload. Now that Amazon's S3 (storage service) is out of beta and EC2 (elastic compute cloud) has added new instance types (the class of machine you can rent) with more CPU and more RAM, I thought it would be interesting to take a look out how their pricing stacks up. The quick conclusion: the m
3 0.18719575 334 high scalability-2008-05-29-Amazon Improves Diagonal Scaling Support with High-CPU Instances
Introduction: Now you can buy more cores on EC2 without adding more machines: The High-CPU Medium Instance is billed at $0.20 (20 cents) per hour. It features 1.7 GB of memory, 5 EC2 Compute Units (2 virtual cores with 2.5 EC2 Compute Units Each), and 350 GB of instance storage, all on a 32-bit platform. The High-CPU Extra Large Instance is billed at $0.80 (80 cents) per hour. It features 7 GB of memory, 20 EC2 Compute Units (8 virtual cores with 2.5 EC2 Compute Units each), and 1,690 GB of instance storage, all on a 64-bit platform. Diagonal Scaling is making a site faster by removing machines. More on this intriguing idea in Diagonal Scaling - Don't Forget to Scale Out AND Up .
4 0.15946111 382 high scalability-2008-09-09-Content Delivery Networks (CDN) – a comprehensive list of providers
Introduction: We build web applications…and there are plenty of them around. Now, if we hit the jackpot and our application becomes very popular, traffic goes up, and our servers are brought down by the hordes of people coming to our website. What do we do in that situation? Of course, I am not talking here about the kind of traffic Digg, Yahoo Buzz or other social media sites can bring to a website, which is temporary overnight traffic, or a website which uses cloud computing like Amazon EC2 service, MediaTemple Grid Service or Mosso Hosting Cloud service. I am talking about traffic that consistently increases over time as the service achieves success. Google.com, Yahoo.com, Myspace.com, Facebook.com, Plentyoffish.com, Linkedin.com, Youtube.com and others are examples of services which have constant high traffic. Knowing that users want speed from their applications, these services will always use a Content Delivery Network (CDN) to deliver that speed. What is a Content Delivery Ne
Introduction: Why isn't Google's aggressive new database pricing strategy getting more pub? That's what Bill Katz , instigator of the GAE Meetup and prize winning science fiction author is wondering: It's surprising that the blogosphere hasn't picked up the biggest difference in pricing: Google's datastore is less than a tenth of the price of Amazon's SimpleDB while offering a better API. If money matters to you then the burn rate under GAE could be convincingly lower. Let's compare the numbers: GAE pricing : * $0.10 - $0.12 per CPU core-hour * $0.15 - $0.18 per GB-month of storage * $0.11 - $0.13 per GB outgoing bandwidth * $0.09 - $0.11 per GB incoming bandwidth SimpleDB Pricing : * $0.14 per Amazon SimpleDB Machine Hour consumed * Structured Data Storage - $1.50 per GB-month * $0.100 per GB - all data transfer in * $0.170 per GB - first 10 TB / month data transfer out (more on the site) Clearly Google priced their services to be competitive
6 0.14473756 1402 high scalability-2013-02-07-Ask HighScalability: Web asset server concept - 3rd party software available?
7 0.14097176 38 high scalability-2007-07-30-Build an Infinitely Scalable Infrastructure for $100 Using Amazon Services
8 0.14077783 619 high scalability-2009-06-05-HotPads Shows the True Cost of Hosting on Amazon
9 0.13374306 1289 high scalability-2012-07-23-State of the CDN: More Traffic, Stable Prices, More Products, Profits - Not So Much
10 0.12894788 283 high scalability-2008-03-18-Shared filesystem on EC2
11 0.12249237 325 high scalability-2008-05-25-How do you explain cloud computing to your grandma?
12 0.11252832 576 high scalability-2009-04-21-What CDN would you recommend?
13 0.10984743 445 high scalability-2008-11-14-Useful Cloud Computing Blogs
14 0.10822405 841 high scalability-2010-06-14-How scalable could be a cPanel Hosting service?
15 0.10538657 888 high scalability-2010-08-27-OpenStack - The Answer to: How do We Compete with Amazon?
17 0.10257564 355 high scalability-2008-07-21-Eucalyptus - Build Your Own Private EC2 Cloud
20 0.10020263 211 high scalability-2008-01-13-Google Reveals New MapReduce Stats
topicId topicWeight
[(0, 0.142), (1, 0.027), (2, 0.018), (3, 0.024), (4, -0.134), (5, -0.06), (6, 0.051), (7, -0.088), (8, 0.082), (9, 0.037), (10, 0.053), (11, -0.136), (12, -0.001), (13, -0.003), (14, -0.017), (15, 0.046), (16, -0.063), (17, 0.024), (18, -0.05), (19, -0.042), (20, -0.028), (21, 0.023), (22, -0.009), (23, 0.071), (24, 0.078), (25, -0.066), (26, 0.026), (27, -0.035), (28, 0.074), (29, -0.005), (30, 0.074), (31, -0.036), (32, -0.041), (33, 0.004), (34, -0.0), (35, 0.043), (36, 0.011), (37, 0.057), (38, 0.007), (39, -0.035), (40, -0.136), (41, -0.036), (42, 0.003), (43, 0.02), (44, -0.094), (45, 0.009), (46, -0.007), (47, 0.003), (48, -0.001), (49, -0.064)]
simIndex simValue blogId blogTitle
same-blog 1 0.97678614 525 high scalability-2009-03-05-Product: Amazon Simple Storage Service
Introduction: Update: HostedFTP.com - Amazon S3 Performance Report . How fast is S3? Based on their own study HostedFTP.com has found: 10 to 12 MB/second when storing and receiving files and 140 ms per file stored as a fixed overhead cost. Update: A Quantitative Comparison of Rackspace and Amazon Cloud Storage Solutions . S3 isn't the only cloud storage service out there. Mosso is saying they can save you so money while offering support. There are number of scenarios in their paper, but For 5TB of cloud storage Mosso will save you 17% over S3 without support and 42% with support. For their CDN on a Global test Mosso says the average response time is 333ms for CloudFront vs. 107ms for Cloud Files which means globally, Cloud Files is 3.1 times or 211% faster than CloudFront. Amazon S3 is storage for the Internet. It is designed to make web-scale computing easier for developers. This service allows you to link directly to files at a cost of 15 cents per GB of storage, and 20 cents per GB
2 0.66309881 195 high scalability-2007-12-28-Amazon's EC2: Pay as You Grow Could Cut Your Costs in Half
Introduction: Update 2: Summize Computes Computing Resources for a Startup . Lots of nice graphs showing Amazon is hard to beat for small machines and become less cost efficient for well used larger machines. Long term storage costs may eat your saving away. And out of cloud bandwidth costs are high. Update: via ProductionScale , a nice Digital Web article on how to setup S3 to store media files and how Blue Origin was able to handle 3.5 million requests and 758 GBs in bandwidth in a single day for very little $$$. Also a Right Scale article on Network performance within Amazon EC2 and to Amazon S3 . 75MB/s between EC2 instances, 10.2MB/s between EC2 and S3 for download, 6.9MB/s upload. Now that Amazon's S3 (storage service) is out of beta and EC2 (elastic compute cloud) has added new instance types (the class of machine you can rent) with more CPU and more RAM, I thought it would be interesting to take a look out how their pricing stacks up. The quick conclusion: the m
3 0.63199872 1053 high scalability-2011-06-06-Apple iCloud: Syncing and Distributed Storage Over Streaming and Centralized Storage
Introduction: There has been a lot of speculation over how Apple's iCloud would work. With the Apple Worldwide Developers Conference keynotes having just completed, we finally learned the truth. We can handle it. They made some interesting and cost effective architecture choices that preserved the value of their devices and the basic model of how their existing services work. A lot of pundits foretold that with all the datacenters Apple was building we would get a streaming music solution. Only one copy of music would be stored and then streamed on demand to everyone. Or they could go the Google brute force method and copy up all a user's music and play it on demand. Apple did neither. The chose an interesting middle path that's not Google, Amazon, or even MobileMe. They key idea is you no longer need a PC. Device content is now synced over the air and is managed by the cloud, not your legacy computer. Your data may not even be stored in the cloud, but the whole management, syncing, and co
4 0.61886269 503 high scalability-2009-01-27-Video: Storage in the Cloud at Joyent
Introduction: Ben Rockwood of Joyent speaks on "Storage in the Cloud" at the first OpenSolaris Storage Summit. Ben is the Director of Systems at Joyent. The Joyent Accelerators are based on OpenSolaris and ZFS. He has deep experience with OpenSolaris in the Real World .
Introduction: Why isn't Google's aggressive new database pricing strategy getting more pub? That's what Bill Katz , instigator of the GAE Meetup and prize winning science fiction author is wondering: It's surprising that the blogosphere hasn't picked up the biggest difference in pricing: Google's datastore is less than a tenth of the price of Amazon's SimpleDB while offering a better API. If money matters to you then the burn rate under GAE could be convincingly lower. Let's compare the numbers: GAE pricing : * $0.10 - $0.12 per CPU core-hour * $0.15 - $0.18 per GB-month of storage * $0.11 - $0.13 per GB outgoing bandwidth * $0.09 - $0.11 per GB incoming bandwidth SimpleDB Pricing : * $0.14 per Amazon SimpleDB Machine Hour consumed * Structured Data Storage - $1.50 per GB-month * $0.100 per GB - all data transfer in * $0.170 per GB - first 10 TB / month data transfer out (more on the site) Clearly Google priced their services to be competitive
6 0.61598521 692 high scalability-2009-09-01-Cheap storage: how backblaze takes matters in hand
7 0.61318022 40 high scalability-2007-07-30-Product: Amazon Elastic Compute Cloud
8 0.60864192 688 high scalability-2009-08-26-Hot Links for 2009-8-26
9 0.60179347 334 high scalability-2008-05-29-Amazon Improves Diagonal Scaling Support with High-CPU Instances
10 0.58128375 43 high scalability-2007-07-30-Product: ImageShack
11 0.56844264 841 high scalability-2010-06-14-How scalable could be a cPanel Hosting service?
12 0.56640178 38 high scalability-2007-07-30-Build an Infinitely Scalable Infrastructure for $100 Using Amazon Services
13 0.55976522 44 high scalability-2007-07-30-Product: Photobucket
14 0.55500752 767 high scalability-2010-01-27-Hot Scalability Links for January 28 2010
15 0.55444837 184 high scalability-2007-12-13-Amazon SimpleDB - Scalable Cloud Database
16 0.55330187 536 high scalability-2009-03-12-Product: Building Next-generation Collaborative cloud-ready applications with Optimus Cloud™
17 0.54725343 1289 high scalability-2012-07-23-State of the CDN: More Traffic, Stable Prices, More Products, Profits - Not So Much
18 0.54470807 355 high scalability-2008-07-21-Eucalyptus - Build Your Own Private EC2 Cloud
19 0.54315984 325 high scalability-2008-05-25-How do you explain cloud computing to your grandma?
20 0.53836018 181 high scalability-2007-12-11-Hosting and CDN for startup video sharing site
topicId topicWeight
[(1, 0.131), (2, 0.21), (61, 0.026), (77, 0.423), (79, 0.07)]
simIndex simValue blogId blogTitle
Introduction: Successful software design is all about trade-offs. In the typical (if there is such a thing) distributed system, recognizing the importance of trade-offs within the design of your architecture is integral to the success of your system. Despite this reality, I see time and time again, developers choosing a particular solution based on an ill-placed belief in their solution as a “silver bullet”, or a solution that conquers all, despite the inevitable occurrence of changing requirements. Regardless of the reasons behind this phenomenon, I’d like to outline a few of the methods I use to ensure that I’m making good scalable decisions without losing sight of the trade-offs that accompany them. I’d also like to compile (pun intended) the issues at hand, by formulating a simple theorem that we can use to describe this oft occurring situation.
Introduction: James Hamilton in Counting Servers is Hard has an awesome breakdown of what one million plus servers really means in terms of resource usage. The summary from his calculations are eye popping: Facilities: 15 to 30 large datacenters Capital expense: $4.25 Billion Total power: 300MW Power Consumption: 2.6TWh annually The power consumption is about the same as used by Nicaragua and the capital cost is about a third of what Americans spent on video games in 2012. Now that's web scale.
3 0.88982981 1116 high scalability-2011-09-15-Paper: It's Time for Low Latency - Inventing the 1 Microsecond Datacenter
Introduction: In It's Time for Low Latency Stephen Rumble et al. explore the idea that it's time to rearchitect our stack to live in the modern era of low-latency datacenter instead of high-latency WANs. The implications for program architectures will be revolutionary . Luiz André Barroso , Distinguished Engineer at Google, sees ultra low latency as a way to make computer resources, to be as much as possible, fungible, that is they are interchangeable and location independent, effectively turning a datacenter into single computer. Abstract from the paper: The operating systems community has ignored network latency for too long. In the past, speed-of-light delays in wide area networks and unoptimized network hardware have made sub-100µs round-trip times impossible. However, in the next few years datacenters will be deployed with low-latency Ethernet. Without the burden of propagation delays in the datacenter campus and network delays in the Ethernet devices, it will be up to us to finish
4 0.86931735 258 high scalability-2008-02-24-Yandex Architecture
Introduction: Update: Anatomy of a crash in a new part of Yandex written in Django . Writing to a magic session variable caused an unexpected write into an InnoDB database on every request. Writes took 6-7 seconds because of index rebuilding. Lots of useful details on the sizing of their system, what went wrong, and how they fixed it. Yandex is a Russian search engine with 3.5 billion pages in their search index. We only know a few fun facts about how they do things, nothing at a detailed architecture level. Hopefully we'll learn more later, but I thought it would still be interesting. From Allen Stern's interview with Yandex's CTO Ilya Segalovich, we learn: 3.5 billion pages in the search index. Over several thousand servers. 35 million searches a day. Several data centers around Russia. Two-layer architecture. The database is split in pieces and when a search is requested, it pulls the bits from the different database servers and brings it together for the user. Languages
5 0.85551274 766 high scalability-2010-01-26-Product: HyperGraphDB - A Graph Database
Introduction: With the success of Neo4j as a graph database in the NoSQL revolution, it's interesting to see another graph database, HyperGraphDB , in the mix. Their quick blurb on HyperGraphDB says it is a: general purpose, extensible, portable, distributed, embeddable, open-source data storage mechanism. It is a graph database designed specifically for artificial intelligence and semantic web projects, it can also be used as an embedded object-oriented database for projects of all sizes. From the NoSQL Archive the summary on HyperGraphDB is: API: Java (and Java Langs), Written in: Java , Query Method: Java or P2P, Replication: P2P , Concurrency: STM , Misc: Open-Source, Especially for AI and Semantic Web. So it has some interesting features, like software transactional memory and P2P for data distribution , but I found that my first and most obvious question was not answered: what the heck is a hypergraph and why do I care? Buried in the tutorial was: A HyperGraphD
6 0.84432608 1195 high scalability-2012-02-17-Stuff The Internet Says On Scalability For February 17, 2012
same-blog 7 0.84385943 525 high scalability-2009-03-05-Product: Amazon Simple Storage Service
8 0.8266452 959 high scalability-2010-12-17-Stuff the Internet Says on Scalability For December 17th, 2010
9 0.8218658 753 high scalability-2009-12-21-Hot Holiday Scalability Links for 2009
10 0.81964928 211 high scalability-2008-01-13-Google Reveals New MapReduce Stats
11 0.81701165 1531 high scalability-2013-10-13-AIDA: Badoo’s journey into Continuous Integration
12 0.78500968 439 high scalability-2008-11-10-Scalability Perspectives #1: Nicholas Carr – The Big Switch
13 0.7614696 1377 high scalability-2012-12-26-Ask HS: What will programming and architecture look like in 2020?
14 0.75585032 612 high scalability-2009-05-31-Parallel Programming for real-world
15 0.75139874 1059 high scalability-2011-06-14-A TripAdvisor Short
16 0.74454373 1158 high scalability-2011-12-16-Stuff The Internet Says On Scalability For December 16, 2011
17 0.74318188 1188 high scalability-2012-02-06-The Design of 99designs - A Clean Tens of Millions Pageviews Architecture
18 0.72706074 212 high scalability-2008-01-14-OpenSpaces.org community site launched - framework for building scale-out applications
19 0.70651096 977 high scalability-2011-01-21-PaaS shouldn’t be built in Silos
20 0.70366061 1571 high scalability-2014-01-02-xkcd: How Standards Proliferate: