Tuesday, May 3, 2011
SummerSlam PPV 2011
How to Make some extra Cash Online - 100% Legit =)
You make 6 cents just for logging in daily.. No limit to withdraw ( withdrawals are done in the form of Paypal or Amazon gift cards )
Compete in contests and meet goals to make even more money $$$ It is possible to make as much as $30+ a day (no limit) Work your own hours, Use a REAL email address when signing up as you will need to confirm (this is to prevent spam / fraud)
Use RTC01 as a referral when signing up for a bonus. (6 cents for daily login)
Click HERE To Sign Up!
Instructional Video:
Let us know your experiences in the comments below / tips, how much you've made, etc =)
Sting Signs WWE Contract 2011

Royal Rumble Winner 2011 Alberto Del Rio
Google uncloaks once-secret server
MOUNTAIN VIEW, Calif.--Google is tight-lipped about its computing operations, but the company for the first time on Wednesday revealed the hardware at the core of its Internet might at a conference here about the increasingly prominent issue of data center efficiency.
Most companies buy servers from the likes of Dell, Hewlett-Packard, IBM, or Sun Microsystems. But Google, which has hundreds of thousands of servers and considers running them part of its core expertise, designs and builds its own. Ben Jai, who designed many of Google's servers, unveiled a modern Google server before the hungry eyes of a technically sophisticated audience.
Google server designer Ben Jai
Google's big surprise: each server has its own 12-volt battery to supply power if there's a problem with the main source of electricity. The company also revealed for the first time that since 2005, its data centers have been composed of standard shipping containers--each with 1,160 servers and a power consumption that can reach 250 kilowatts.
It may sound geeky, but a number of attendees--the kind of folks who run data centers packed with thousands of servers for a living--were surprised not only by Google's built-in battery approach, but by the fact that the company has kept it secret for years. Jai said in an interview that Google has been using the design since 2005 and now is in its sixth or seventh generation of design.
"It was our Manhattan Project," Jai said of the design.
Google has an obsessive focus on energy efficiency and now is sharing more of its experience with the world. With the recession pressuring operations budgets, environmental concerns waxing, and energy prices and constraints increasing, the time is ripe for Google to do more efficiency evangelism, said Urs Hoelzle, Google's vice president of operations.
"There wasn't much benefit in trying to preach if people weren't interested in it," said Hoelzle, but now attitudes have changed.
The company also focuses on data center issues such as power distribution, cooling, and ensuring hot and cool air don't intermingle, said Chris Malone, who's involved in the data center design and efficiency measurement. Google's data centers now have reached efficiency levels that the Environmental Protection Agency hopes will be attainable in 2011 using advanced technology.
"We've achieved this now by application of best practices and some innovations--nothing really inaccessible to the rest of the market," Malone said.
The rear side of Google's server.
Why built-in batteries?
Why is the battery approach significant? Money.
Typical data centers rely on large, centralized machines called uninterruptible power supplies (UPS)--essentially giant batteries that kick in when the main supply fails and before generators have time to kick in. Building the power supply into the server is cheaper and means costs are matched directly to the number of servers, Jai said.
"This is much cheaper than huge centralized UPS," he said. "Therefore no wasted capacity."
Efficiency is another financial factor. Large UPSs can reach 92 to 95 percent efficiency, meaning that a large amount of power is squandered. The server-mounted batteries do better, Jai said: "We were able to measure our actual usage to greater than 99.9 percent efficiency."
Urs Hoelzle, Google's vice president of operations
The Google server was 3.5 inches thick--2U, or 2 rack units, in data center parlance. It had two processors, two hard drives, and eight memory slots mounted on a motherboard built by Gigabyte. Google uses x86 processors from both AMD and Intel, Jai said, and Google uses the battery design on its network equipment, too.
Efficiency is important not just because improving it cuts power consumption costs, but also because inefficiencies typically produce waste heat that requires yet more expense in cooling.
Costs add up
Google operates servers at a tremendous scale, and these costs add up quickly.
Jai has borne a lot of the burden himself. He was the only electrical engineer on the server design job from 2003 to 2005, he said. "I worked 14-hour days for two and a half years," he said, before more employees were hired to share the work.
Google has patents on the built-in battery design, "but I think we'd be willing to license them to vendors," Hoelzle said.
Another illustration of Google's obsession with efficiency comes through power supply design. Power supplies convert conventional AC (alternating current--what you get from a wall socket) electricity into the DC (direct current--what you get from a battery) electricity, and typical power supplies provide computers with both 5-volt and 12-volt DC power. Google's designs supply only 12-volt power, with the necessary conversions taking place on the motherboard.
Google's data center efficiency has been improving gradually.
That adds $1 or $2 to the cost of the motherboard, but it's worth it not just because the power supply is cheaper, but because the power supply can be run closer to its peak capacity, which means it runs much more efficiently. Google even pays attention to the greater efficiency of transmitting power over copper wires at 12 volts compared to 5 volts.
Google also revealed new performance results for data center energy efficiency measured by a standard called power usage effectiveness. PUE, developed by a consortium called the Green Grid, measures how much power goes directly to computing compared to ancillary services such as lighting and cooling. A perfect score of 1 means no power goes to the extra costs; 1.5 means that ancillary services consume half the power devoted to computing.
Google's PUE scores are enviably low, but the company is working to lower them further. In the third quarter of 2008, Google's PUE was 1.21, but it dropped to 1.20 for the fourth quarter and to 1.19 for the first quarter of 2009 through March 15, Malone said.
Older Google facilities generally have higher PUEs, he said; the best has a score of 1.12. When the weather gets warmer, Google notices is that it's harder to keep servers cool.
An excerpt from a video tour Google presented of its data center containers. Like conventional data centers, Google's shipping containers have raised floors.
Shipping containers
Most people buy computers one at a time, but Google thinks on a very different scale. Jimmy Clidaras revealed that the core of the company's data centers are composed of standard 1AAA shipping containers packed with 1,160 servers each, with many containers in each data center.
Modular data centers are not unique to Google; Sun Microsystems and Rackable Systems both sell them. But Google started using them in 2005.
Google's first experiments had some rough patches, though, Clidaras said--for example when they found the first crane they used wasn't big enough to actually lift one.
Overall, Google's choices have been driven by a broad analysis on cost that encompasses software, hardware, and facilities.
"Early on, there was an emphasis on the dollar per (search) query," Hoelzle said. "We were forced to focus. Revenue per query is very low."
Mainstream servers with x86 processors were the only option, he added. "Ten years ago...it was clear the only way to make (search) work as free product was to run on relatively cheap hardware. You can't run it on a mainframe. The margins just don't work out," he said.
Operating at Google's scale has its challenges, but it also has its silver linings. For example, a given investment on research can be applied to a larger amount of infrastructure, yielding return faster, Hoelzle said.
Read more and view the design: http://news.cnet.com/8301-1001_3-10209580-92.html
Hotmail Passwords Leak
Details of a Hotmail security breach emerged early yesterday, with approximately at least 10,000 confirmed accounts having been compromised.
While the exact number of accounts has not yet been confirmed, the breach was likely the result of a phishing campaign -- a different kind of hack that uses fake web pages in order to acquire all kinds of sensitive information, from login data and passwords (such as in this case) to credit card and social security numbers. (Source: cnet.com)
"We determined that this was not a breach of internal Microsoft data and initiated our standard process of working to help customers regain control of their accounts," said a Microsoft representative through email to Computerworld. (Source: computerworld.com)
"Over the weekend Microsoft learned that several thousand Windows Live Hotmail customers' credentials were exposed on a third-party site due to a likely phishing scheme," the spokesperson admitted.
Update: Microsoft blocked access to all accounts that were listed as compromised. So for all those worried if their account was on the exposed list. If you were able to login normally with your account then your login details weren't exposed. I would recommend everyone to change their passwords after logging in in the case that there were more accounts compromised that weren't listed.
Office Design Gallery
MR Design Office by Schemata Architecture Office





Source http://dezeen.com
Red Bull Australia Headquarters




Red Bull Australia Headquarters designed by Sheargold
Photos by Tyrone Branigan
Wikileaks Host Bahnhof AB



By Julian Masters http://www.julianmasters.com.au
Cheap Monday Office





Fashion brand Cheap Monday studio office by Swedish designers Uglycute
Photos are by Mikael Olsson
United Tannery and Boot Factory



Photos by Derek Swalwell
Design by Wolveridge Architects
Konstruplus Studio by MMASA Architects




Source: http://www.mmasa.net
Vodafone Portugal




Photography by ultimasreportagens.com
designed by architects José António Barbosa and Pedro Guimarães of Barbosa Guimarães Arquitectos.
Source http://thecoolhunter.net




Source: http://trendland.com
Candy Black studio




Thanks to Fabien from http://www.graphic-exchange.com/home.html for sending these in. The big graphic can also be seen here: http://www.harmonie-interieure.com/possible
Dentsu London





Design by http://www.essentiadesigns.com/
Skinny Corp





Source: http://www.avidos.net/detalles




Designer: One Work Place http://www.oneworkplace.com
D1 Creative


Designer: The Architecture Office http://www.archoffice.co.nz
Fog Creek


Designer: Roy Leone http://www.royleone.com
Source: Joel On Software http://www.joelonsoftware.com/articles/BionicOffice.html
Mabeg


Designer: Grimshaw Architects http://www.ngrimshaw.co.uk
Push Orlando



Designer: The Evans Group http://www.theevansgroup.com
Lumina Americas



Designer: F2 Inc http://www.f2inc.com/
Inklude




Source: http://www.avidos.net