Top 3 News items fromÂ scalebig
- The big news is that Amazon has got into PAAS big time. I predicted it only a couple of days ago ( I said they will launch within next 1 year ). With beanstalk they plan to provide containers into which users can upload code to and let AWS manage rest of the complexities of things around it. They are starting with a tomcat based container for now and have mentioned plans to build other containers. Read more about it at “All things distributed“
- As weird as would sound, GoGrid is building the private cloud over public infrastructure. They are doing this just to let let CIO claim that they own the servers. This allows CIOs claim to be on two boat at the same time. At some point though CIOs will have to make a call and abandon one. BTW, this is not very different from managed infrastructure, with the exception that there now exists a virtualization toolkit to manage VMs on this managed infrastructure.
- Hbase 0.90.0 is released. Lots of interesting improvements which a lot of people were waiting for. Alex has some observations here.
I found the top two stories on scalebig last night to be interesting enough for me to dig a little deeper. The one which surprised me the most was William Vambenepeâ€™s post about why he thinks that REST APIs doesnâ€™t matter in context of cloud management. While REST might be ideal for many different things, including web based applications which are accessed mostly by the browsers, Amazon chose to avoid REST for most of its infrastructure management APIs.
Has this lack of REStfulness stopped anyone from using it? Has it limited the scale of systems deployed on AWS? Does it limit the flexibility of the Cloud offering and somehow force people to consume more resources than they need? Has it made the Amazon Cloud less secure? Has it restricted the scope of platforms and languages from which the API can be invoked? Does it require more experienced engineers than competing solutions?
I donâ€™t see any sign that the answer is â€œyesâ€ to any of these questions. Considering the scale of the service, it would be a multi-million dollars blunder if indeed one of them had a positive answer.
Hereâ€™s a rule of thumb. If most invocations of your API come via libraries for object-oriented languages that more or less map each HTTP request to a method call, it probably doesnâ€™t matter very much how RESTful your API is.
The Rackspace people are technically right when they point out the benefits of their API compared to Amazonâ€™s. But itâ€™s a rounding error compared to the innovation, pragmatism and frequency of iteration that distinguishes the services provided by Amazon. Itâ€™s the content that matters.
And the other big news was of course the launch of a new cloud datastore by salesforce at Database.com. Interestingly, you should notice, that they decided to brand it with its own website instead of making it part of its existing set of services. Its possible they did it to distance this new service from an impression that its only useful for applications which need other salesforce services. For more in-depth technical information continue reading here.
The infrastructure promises automatic tuning, upgrades, backups and replication to remote data centers, and automatic creation of sandboxes for development, test and training. Database.com offers enterprise search services, allowing developers to access a full-text search engine that respects enterprise security rules
In terms of pricing, Database.com access will be free for 3 users, and up to 100,000 records and 50,000 transactions per month. The platform will $10 per month for each set of 100,000 records beyond that and another $10 per month for each set of 150,000 transactions beyond that benchmark. The enterprise-level services will be an additional $10 per user per month and will include user identity, authentication and row-level security access controls.
Other references: Dreamforce 2010 â€“ Database.com Launch Interview with Eric Stahl Read more
The complete announcement is here, but here are the changes for the java SDK. The two big changes I liked is the fact that there is now an â€œalways onâ€ feature, and â€œtasksâ€ feature has graduated out of beta/testing.
- The Always On feature allows applications to pay and keep 3 instances of
their application always running, which can significantly reduce application latency.
- Developers can now enable Warmup Requests. By specifying a handler in an app’s appengine-web.xml, App Engine will attempt to to send a Warmup Request to initialize new instances before a user interacts with it. This can reduce the latency an end-user sees for initializing your application.
- The Channel API is now available for all users.
- Task Queue has been officially released, and is no longer an experimental feature. The API import paths that use ‘labs’ have been deprecated. Task queue storage will count towards an application’s overall storage quota, and will thus be charged for.
- The deadline for Task Queue and Cron requests has been raised to 10 minutes. Datastore and API deadlines within those requests remain unchanged.
- For the Task Queue, developers can specify task retry-parameters in their queue.xml.
- Metadata Queries on the datastore for datastore kinds, namespaces, and entity properties are available.
- URL Fetch allowed response size has been increased, up to 32 MB. Request
size is still limited to 1 MB.
- The Admin Console Blacklist page lists the top blacklist rejected visitors.
- The automatic image thumbnailing service supports arbitrary crop sizes up to 1600px.
- Overall average instance latency in the Admin Console is now a weighted average over QPS per instance.
- Added a low-level AysncDatastoreService for making calls to the datastore asynchronously.
- Added a getBodyAsBytes() method to QueueStateInfo.TaskStateInfo, this returns the body of the task state as a pure byte-string.
- The whitelist has been updated to include all classes from javax.xml.soap.
- Fixed an issue sending email to multiple recipients. http://code.google.com/p/googleappengine/issues/detail?id=1623
I have some experience with Akamaiâ€™s WAA (Web applications archive) service, which Iâ€™ve been using in my professional capacity for a few years now. And Iâ€™ve have been curious about how cloudfront compares with it. Until a few weeks ago, Cloudfront didnâ€™t have a key feature which I think was critical for it to win the traditional CDN customers. â€œCustom originâ€ is an amazing new feature which I finally got to test last night and here are my notes for those who are curious as well.
My test application which I tried to convert was my news aggregator portal http://www.scalebig.com/. The application consists of a rapidly changing front page (few times a day) , a collection of old pages archived in a sub directory and some other webpage elements like headers, footers, images, style-sheets etc.
- While Amazon Coudfront does have a presence on AWS management console, it only supports S3 buckets as origins.
- Since my application didnâ€™t have any components which requires server side processing, I tried to put the whole website on an S3 bucket and tried to use S3 as the origin.
- When I initially set it up, I ended up with multiple URLs which I had to understand
- S3 URL â€“ This is the unique URL to your S3 bucket. All requests to this URL will go to Amazons S3 server cluster, and if your objects are marked as private, anyone can get these objects. The object could be a movie, an image, or even an HTML file.
- Cloudfront URL â€“ This is the unique Cloudfront URL which maps to your S3 resource through the cloudfront network. For all practical purposes its the same as the first one, except that this is through the CDN service.
- Your own domain name â€“ This is the actual URL which end users will see, which will be a CNAME to the cloudfront URL.
- So in my case, I configured the DNS entry for www.scalebig.com to point to DNS entry Cloudfront service created for me (dbnqedizktbfa.cloudfront.net).
- First thing which broke is that I forgot that this is just an S3 bucket, so it canâ€™t handle things like â€œsparsed htmlâ€ to dynamically append headers/footers. I also realized that it canâ€™t control cache policies, setup expiry, etc. But the worst problem was that if you went to â€œhttp://www.scalebig.com/â€ it would throw an error. It was expecting a file name, so http://www.scalebig.com/index.html would have worked.
- In short I realized that my idea of using S3 as a webserver full of holes.
- When I started digging for options to enable â€œcustom originâ€ I realized that those options do not exist on the AWS management console !!. I was instead directed to some third party applications to do this instead. (most of them were commercial products, except two)
- I finally created the cloudfront configuration using Cloudberry S3 Explorer PRO which allowed me to point Cloudfront to a custom domain name (instead of an S3 resource).
- In my case my server was running on EC2 with a public reserved IP. Iâ€™m not yet using AWS ELB (Elastic loadbalancer).
- Once I got that working, which literally worked out of the box, the next challenge is to setup the cache controls and expiries working. If they are set incorrectly, it may stop users from getting latest content. I setup the policies using â€œ.htaccessâ€. Below Iâ€™ve attached a part of the .htaccess I have for the /index.html page which is updated many times a day. There is a similar .htaccess page for rest of the website which recommends a much longer expiry.
- Finally I realized that it is possible that I might have to invalidate parts of the caches at times (could be due to a bug). Cloudberry and AWS management console didnâ€™t have any option avaliable, but apparently â€œbotoâ€ has some APIs which can work with Amazon cloudfront APIs to do this.
# turn on the module for this directory
# set default
ExpiresDefault "access plus 1 hours"
ExpiresByType image/jpg "access plus 1 hours"
ExpiresByType image/gif "access plus 1 hours"
ExpiresByType image/jpeg "access plus 1 hours"
ExpiresByType image/png "access plus 1 hours"
ExpiresByType text/css "access plus 1 hours"
ExpiresByType application/x-shockwave-flash "access plus 1 hours"
Header set Cache-Control "max-age=3600"
Here is how I would summarize the current state of Amazon cloudfront.
- Its definitely ready for static websites which donâ€™t have any server side execution code.
- Cloudfront only accepts GET and HEAD requests
- Cloudfront ignores cookies, so server canâ€™t set any. (Browser based cookie management will still work, which could be used to keep in-browser session data)
- While Cloudfront can log access logs to an S3 bucket of your choice, Iâ€™ll recommend using something like Google Analytics to do log analysis.
- Iâ€™ll recommend buying one of the commercial third party products if you want to use Custom Origin and would recommend reading more about the protocols/APIs before you fully trust a production service to Cloudfront.
- I wish Cloudfront starts supporting something like ESI, which could effectively make an S3 bucket a full fledged webserver without the need of having a running EC2 instance all the time.
- Overall Cloudfront has a very long way to go, in the number of features, to be treated as a competitor for Akamaiâ€™s current range of services.
- And if you look at Akamaiâ€™s current world wide presence, Cloudfront is just a tiny blip. [ Cloudfront edge locations ]
- But I suspect that Cloudfrontâ€™s continuous evolution is being watched by many and the next set of features could change the balance.
Iâ€™m planning to leave http://www.scalebig.com/ on Cloudfront for some time to learn a little more about its operational issues. If you have been using Cloudfront please feel free to leave comments about what important features, you think, are still missing.
Any blog which promotes the concept of cloud infrastructure would be doing injustice if it doesnâ€™t provide references to implementations where it failed horribly. Here is an excellent post by Carlos Ble where he lists out all the problems he faced on Google App engine (python). He lists 13 different limitations, most of which are very well known facts, and then lists some more frustrating reasons why he had to dump the solution and look for an alternative.
The tone of the voice is understandable, and while it might look like App-Engine-bashing, I see it as a great story which others could lean from.
For us, GAE has been a failure like Wave or Buzz were but this time, we have paid it with our money. I’ve been too stubborn just because this great company was behind the platform but I’ve learned an important lesson: good companies make mistakes too. I didn’t do enough spikes before developing actual features. I should have performed more proofs of concept before investing so much money. I was blind.
Cloud is not for everyone or for all problems. While some of these technologies take away your growing pain points, they assume you are ok with some of the limitations. If you were surprised by these limitations after you are neck deep in coding, then you didnâ€™t do your homework.
Here are the 13 points issues he pointed out. I havenâ€™t used Google App engine lately, but my understanding is that App engine team have solved, or on the path of solving (or reducing pain) some of these issues.
- Requires Python 2.5
- Cant use HTTPS
- 30 seconds to run
- URL fetch gets only 5 seconds
- Canâ€™t use python libraries compiled in C
- No â€œLIKEâ€ operators in datastore
- Canâ€™t join tables
- â€œToo many indexesâ€
- Only 1000 records at a time returned
- Datastore and memcache can fail at times
- Max memcache size is 1MB
While some of the interest in moving towards public cloud is based on sound economics, there is a small segment of this movement purely due to the â€œherd mentalityâ€.
The slide on the right is from a Microsoft publication shows that larger networks may be less economical on the cloud (at least today).
Richard Farley, has been discussing this very topic for few months now. He observed that a medium sized organization which already has a decent IT infrastructure including a dedicated IT staff to support it has a significantly smaller overhead than what cloud vendors might make it look like.
Here is a small snippet from his blog. If you are not afraid to get dirty with numbers read the rest here.
Now, we know we need 300 virtual servers, each of which consumes 12.5% of a physical host. This means we need a total of 37.5 physical hosts. Our vendor tells us these servers can be had for $7k each including tax and delivery with the cabinet. We canâ€™t buy a half server, and want to have an extra server on hand in case one breaks. This brings our total to 39 at a cost of $273k. Adding in the cost of the cabinet, weâ€™re up to $300k.
There are several non-capital costs we now have to factor in. Your vendor will provide warranty, support and on-site hardware replacement service for the cabinet and servers for $15k per year. Figure you will need to allocate around 5% of the time of one of your sys admins to deal with hardware issues (i.e., coordinating repairs with the vendor) at a cost of around $8k per year in salary and benefits. Figure power and cooling for the cabinet will also cost $12k per year. In total, your non-capital yearly costs add up to $35k.
One thing which posts doesnâ€™t clearly articulate, is the fact that while long term infrastructure is cheaper to host in private cloud, its may still be more economical to use public cloud for short term resource intensive projects.
Summary from Jamesâ€™ keynote talk at Velocity 2010
- Pace of Innovation â€“ Datacenter pace of innovation is increasing. The high focus on infrastructure innovation is driving down the cost, increasing reliability and reducing resource consumption which ultimate drives down cost.
- Where does the money go ?
- 54% on servers, 8% on networking, 21% on power distribution, 13% on power, 5% on other infrastructure requirements
- 34% costs related to power
- Cost of power is trending up
- Clouds efficiency â€“ server utilization in our industry is around 10 to 15% range
- Avoid holes in the infrastructure use
- Break jobs into smaller chunks, queue them where ever possible
- Power distribution â€“ 11 to 12% lost in distribution
- Rules to minimize power distribution losses
- Oversell power â€“ setup more servers than power available. 100% of servers never required in a regular datacenter.
- Avoid voltage conversions
- Increase efficiency of conversions
- High voltage as close to load as possible
- Size voltage regulators to load and use efficient parts
- High voltage direct current a small potential gain
- Mechanical Systems â€“ One of the biggest saving is in cooling
- What parts are involved ? – Cooling tower, heat exchanges, pumps, evaporators, compressors, condensers, pumpsâ€¦ and so on.
- Efficiency of these systems and power required to get this done depends on the difference in the desired temperature and the current room temperature
- Separate hot and cold islesâ€¦ insulate them (donâ€™t break the fire codes)
- Increase the operating temperature of servers
- Most are between 61 and 84
- Telco standard is 104F (Game consoles are even higher)
- Limiting factors to high temp operation
- Higher fan power trade-off
- More semiconductor leakage current
- Possible negative failure rate impact
- Avoid direct expansion cooling entirely
- Air side economization
- Higher data center temperature
- Evaporative cooling
- Requires filtration
- Particulate and chemical pollution
- Networking gear
- Current networks are over-subscribed
- Forces workload placement restrictions
- Goal: all points in datacenter equidistant.
- Mainframe model goes commodity
- Competition at each layer rather than vertical integration
- Openflow: open S/W platform
- Distributed control plane to central control
Yesterday Google formally announced Google Storage to a few (5000?) of us at Google I/O. Here is the gist of this as I see it from the various discussions/talks I attended.
To begin with, I have to point out that there is almost nothing new in what Google has proposed to provide. Amazon has been doing this for years with its S3. The key difference is that if you are a google customer you wonâ€™t have to look elsewhere for storage services like this one.
Lets get the technical details out
- Its tries to implement a Strong consistency model (CA of the CAP: Consistent and Available). Which means data you store is automatically replicated in a consistent way across multiple datacenter
- Currently it replicates to multiple locations within US. In future it does plan to replicate across continents.
- Currently there are no controls to control how replication happens or to where. They plan to learn from usage in beta period and develop controls over time.
- There are two basic building blocks for objects
- Buckets â€“ Containers
All objects are stored in flat container. However, the tools understand â€œ/â€ and â€œ*â€ (wild cards) and does the right thing when used correctly
- Objects â€“ objects/files inside those containers
- Implements RESTful APIs (GET/PUT/POST/DELETE/HEAD/etc)
- All resources are identified by a URI
- No theoretical size limit of Buckets or containers. However a 100GB limit per account would be imposed during beta phase.
- Its of course built on Google very well tested, scalable, highly available infrastructure
- It provides multiple, flexible authentication and sharing models
- Does support standard public/private key based auth
- Will also have integration with some kind of groups which will allow object to be shared with or controlled by with multiple identities.
- ACLs can be applied to both Buckets and Objects
- Control who can list objects
- Who can create/delete objects
- Who can read/write into the bucket
- Who can read
- Who can read/write
- There were two tools mentioned during the talk
- GS manager looks like a web application which allows an admin to manage this service
- GS util is more like the shell tools AWS provides for S3.
- As I mentioned before GS util accepts wild card
- So something like this is possible
- gsutil cp gs://gs2010/* /home/rkt/gs2010
- The service was created with â€œdata liberationâ€ as one of the goals. As shown by the previous command it takes just one line of code to transfer all of your data out.
- Resume feature (if connection breaks during a big upload) is not available yet, but thats on the roadmap.
- Groups feature was discussed a lot, but its not ready in the current release
- Versioning feature is not available. Wasnâ€™t clear if its on the roadmap or how long before its implemented.
Few other notes.
- Its not clear how this plays with the â€œstorage serviceâ€ Google currently provides for Gmail/Docs storage. From what I heard this is not related to that storage service at all and there are no plans to integrate it.
- The service is free in beta period to all developers who get access to it, but when its released it will follow a pricing model similar others in the industry. The pricing model is already published on their website
- The speakers and the product managers didnâ€™t comment on whether storage access from google apps engine would be charged (or at what rate)
- They do provide MD5 signatures as a way of verifying if an object on the client is same as the object on the server, but its not used storing files itself. (So MD5 collisions issue shouldnâ€™t be a problem)
- US Navy is already using this service with about 80TB of data on Google Storage, and from what I heard they looked pretty happy talking about it.
I suspect this product will be in beta for a while before they release it out in the open.
Ted Dziuba has a post about â€œI canâ€™t wait for NoSQL to Dieâ€. The basic argument he makes is that one has to be at the size Google is to really benefit from NoSQL. I think he is missing the point.
Here are my observations.
- This is similar to the argument the traditional DB vendors were making when companies started switching away from the likes of Oracle/DB2 to MySQL. The difference between then and now is that before it was Large established databases vendors against the smaller (open-source) ones, and now its RDBMS vs non-RDBMS datastores.
- Why NoSQL: The biggest difference between an RDBMS and a NoSQL datastore is the fact that NoSQL datastructures have no pre-defined schemas. That doesnâ€™t mean that the developers donâ€™t have to think about the data structure before using a NoSQL solution, but it does provide the opportunity to developers to add new columns which were not thought of at design time with little or no impact on applications using it. You can add and remove columns on the fly on most RDBMS as well, but those changes are usually considered significant. Also keep in mind that while NoSQL datastores could add columns at the row level, RDBMS solutions can only do it at the table level.
- Scalability: There are basically two ways to scale any web application.
- The first way is to build the app and leave the scalability issues for later (let the DBAs to figure out). This is an expensive iterative process which takes time to perfect. The issues around scalability and availability could be so complex that one may not be able to predict all the issues until they get used in production.
- The second way is to train the programmers to architect the database so that it can scale better once it hits production. There is a significant upfront cost, but it pays over time.
- NoSQL is the third way of doing it.
- It restricts programmers by allowing only those operations and data-structures which can scale
- And programmers who manage to figure out how to use it, have found that the these kind of restrictions guarantee significantly higher horizontal scalability than traditional RDBMS.
- By architecting databases before the product is launched, it also reduces the amount of outage and post-deployment migrations.
- High Availability: NoSQL is not just about scalability. Its also about â€œhigh-availabilityâ€ at a cheaper cost.
- While Ted did mention that some of the operations in Cassandra requires a restart, he forgot to mention that it doesnâ€™t require all the nodes to be restarted at the same time. The cassandra datastore continues to be available even without many of its nodes. This is a common theme across most of the NoSQL based datastores. [CASSANDRA-44]
- High availability over long distances with flaky network connection is not trivial to implement using traditional RDBMS based databases.
- You donâ€™t have to be Google to see benefits of using NoSQL.
- If you are using S3 or SimpleDB on AWS or using datastores on Googleâ€™s Appengine then you are already using NoSQL. Many of the smaller startups are actually finding AWS/GAE to be cheaper than hosting their own servers.
- One can still chose to use RDS like RDBMS solution, but they donâ€™t get the benefit of high-availability and scalability which S3/SimpleDB offers out-of-the-box.
- While scalability to terabytes may not be a requirement for many of the smaller organizations, high availability is absolutely essential for most organizations today. RDBMS based solutions can do that, but setting up multi-master replication across two datacenters is non-trivial
- Migration from RDBMS to NoSQL is not simple: I think Ted is right that not everyone will have success in cutting over from RDBMS to non-RDBMS world in one weekend. The reports of websites switching over to NoSQL overnight is sometimes grossly exaggerated. Most of these companies have been working on this for months if not years. And they would do extensive scalability, performance, availability and disaster-recovery tests before they put it in production.
- RDBMS is not going anywhere: I also agree with Ted that RDBMS is not going anywhere anytime soon. Especially in organizations which are already using it. In fact most NoSQL datastores still havenâ€™t figured out how to implement the level of security traditional RDBMS provide. I think thats the core reason why Google is still using it for some of its operational needs.
Finally, its my personal opinion that â€œCloud computingâ€ and commoditization of storage and servers were the key catalysts for the launch of so many NoSQL implementations. The ability to control infrastructure with APIs was a huge incentive for the developers to develop datastores which could scale dynamically as well. While Oracle/MySQL are not going anywhere anytime soon, â€œNoSQLâ€ movement is definitely here to stay and I wonâ€™t be surprised if it evolves more on the way.
- Haters Gonna Hate
- Reddit: learning from mistakes
- Digg: Saying yes to NoSQL; Going steady with Cassandra
- Twitter @ 2009/07 : Up and running with cassandra
- Twitter @ 2010/03 : Ryan King about Twitter and Cassandra
- NoSQL vs RDBMS: Let the flames begin !
- Brewerâ€™s CAP theorem on Distributed systems
- Database scalability
- What is scalability ?
- Thoughts on NoSQL
We discussed Brewerâ€™s Theorm a few days ago and how its challenging to obtain Consistency, Availability and Partition tolerance in any distributed system. We also discussed that many of the distributed datastores allow CAP to be tweaked to attain certain operational goals.
Amazon SimpleDB, which was released as an â€œEventually Consistentâ€ datastore, today launched a few features to do just that.
- Consistent reads: Select and GetAttributes request now include an optional Boolean flag â€œConsistentReadâ€ which requests datastore to return consistent results only. If you have noticed scenarios where read right after a write returned an old value, it shouldnâ€™t happen anymore.
- Conditional put/puts, delete/deletes : By providing â€œconditionsâ€ in the form of a key/value pair SimpleDB can now conditionally execute/discard an operation. This might look like a minor feature, but can go a long way in providing reliable datastore operations.
Even though SimpleDB now enables operations that support a stronger consistency model, under the covers SimpleDB remains the same highly-scalable, highly-available, and highly durable structured data store. Even under extreme failure scenarios, such as complete datacenter failures, SimpleDB is architected to continue to operate reliably. However when one of these extreme failure conditions occurs it may be that the stronger consistency options are briefly not available while the software reorganizes itself to ensure that it can provide strong consistency. Under those conditions the default, eventually consistent read will remain available to use.