It is a rare criticism of elite American university students that they do not think big enough. But that is exactly the complaint from some of the largest technology companies and the federal government.
At the heart of this criticism is data. Researchers and workers in fields as diverse as biotechnology, astronomy and computer science will soon find themselves overwhelmed with information. Better telescopes and genome sequencers are as much to blame for this data glut as are faster computers and bigger hard drives.
While consumers are just starting to comprehend the idea of buying external hard drives for the home capable of storing a terabyte of data, computer scientists need to grapple with data sets thousands of times as large and growing ever larger. (A single terabyte equals 1,000 gigabytes and could store about 1,000 copies of the Encyclopedia Britannica.)
The next generation of computer scientists has to think in terms of what could be described as Internet scale. Facebook, for example, uses more than 1 petabyte of storage space to manage its users’ 40 billion photos. (A petabyte is about 1,000 times as large as a terabyte, and could store about 500 billion pages of text.)
It was not long ago that the notion of one company having anything close to 40 billion photos would have seemed tough to fathom. Google, meanwhile, churns through 20 times that amount of information every single day just running data analysis jobs. In short order, DNA sequencing systems too will generate many petabytes of information a year.
“It sounds like science fiction, but soon enough, you’ll hand a machine a strand of hair, and a DNA sequence will come out the other side,” said Jimmy Lin, an associate professor at the University of Maryland, during a technology conference last week.
The big question is whether the person on the other side of that machine will have the wherewithal to do something interesting with an almost limitless supply of genetic information.
At the moment, companies like IBM and Google have their doubts.
For the most part, university students have used rather modest computing systems to support their studies. They are learning to collect and manipulate information on personal computers or what are known as clusters, where computer servers are cabled together to form a larger computer. But even these machines fail to churn through enough data to really challenge and train a young mind meant to ponder the mega-scale problems of tomorrow.
“If they imprint on these small systems, that becomes their frame of reference and what they’re always thinking about,” said Jim Spohrer, a director at IBM’s Almaden Research Center.
Two years ago, IBM and Google set out to change the mindset at universities by giving students broad access to some of the largest computers on the planet. The companies then outfitted the computers with software that Internet companies use to tackle their toughest data analysis jobs.
And, rather than building a big computer at each university, the companies created a system that let students and researchers tap into giant computers over the Internet.
This year, the National Science Foundation, a federal government agency, issued a vote of confidence for the project by splitting US$5 million among 14 universities that want to teach their students how to grapple with big data questions.
The types of projects the 14 universities have already tackled veer into the mind-bending.
For example, Andrew Connolly, an associate professor at the University of Washington, has turned to the high-powered computers to aid his work on the evolution of galaxies. Connolly works with data gathered by large telescopes that inch their way across the sky taking pictures of various objects.
The largest public database of such images available today comes from the Sloan Digital Sky Survey, which has about 80 terabytes of data, Connolly said. A new system called the Large Synoptic Survey Telescope is set to take more detailed images of larger chunks of the sky and produce about 30 terabytes of data each night. Connolly’s graduate students have been set to work trying to figure out ways of coping with this much information.
Purdue looks to carry out techniques used to map the interactions between people in social networks into the biological realm. Researchers are creating complex diagrams that illuminate the links between chemical reactions taking place in cells.
A similar effort at the University of California, Santa Barbara, centers on making a simple software interface — akin to the Google search bar — that will let researchers examine huge biological data sets for answers to specific queries.
Lin has encouraged his students to illuminate data with the help of Hadoop, an open-source software package that companies like Facebook and Yahoo use to split vast amounts of information into more manageable chunks.
One of these projects included a deep dive into the reams of documents released after the government’s probe into Enron, to create an analysis system that could identify how one employee’s internal communications had been connected to those from other employees and who had originated a specific decision.
Lin shares the opinion of numerous other researchers that learning these types of analysis techniques will be vital for students in the coming years.
“Science these days has basically turned into a data-management problem,” Lin said.
By donating their computing wares to the universities, Google and IBM hope to train a new breed of engineers and scientists to think in Internet scale.
Of course, it’s not all good will backing these gestures. IBM is looking for big data experts that can complement its consulting in areas like health care and financial services. It has already started working with customers to put together analytics systems built on top of Hadoop. Meanwhile, Google promotes just about anything that creates more information to index and search.
Nonetheless, the universities and the government benefit from IBM and Google providing access to big data sets for experiments, simpler software and their computing wares.
“Historically, it has been tough to get the type of data these researchers need out of industry,” said James French, a research director at the National Science Foundation.
“But we’re at this point where a biologist needs to see these types of volumes of information to begin to think about what is possible in terms of commercial applications,” he said.
US President Donald Trump and Chinese President Xi Jinping (習近平) were born under the sign of Gemini. Geminis are known for their intelligence, creativity, adaptability and flexibility. It is unlikely, then, that the trade conflict between the US and China would escalate into a catastrophic collision. It is more probable that both sides would seek a way to de-escalate, paving the way for a Trump-Xi summit that allows the global economy some breathing room. Practically speaking, China and the US have vulnerabilities, and a prolonged trade war would be damaging for both. In the US, the electoral system means that public opinion
They did it again. For the whole world to see: an image of a Taiwan flag crushed by an industrial press, and the horrifying warning that “it’s closer than you think.” All with the seal of authenticity that only a reputable international media outlet can give. The Economist turned what looks like a pastiche of a poster for a grim horror movie into a truth everyone can digest, accept, and use to support exactly the opinion China wants you to have: It is over and done, Taiwan is doomed. Four years after inaccurately naming Taiwan the most dangerous place on
Wherever one looks, the United States is ceding ground to China. From foreign aid to foreign trade, and from reorganizations to organizational guidance, the Trump administration has embarked on a stunning effort to hobble itself in grappling with what his own secretary of state calls “the most potent and dangerous near-peer adversary this nation has ever confronted.” The problems start at the Department of State. Secretary of State Marco Rubio has asserted that “it’s not normal for the world to simply have a unipolar power” and that the world has returned to multipolarity, with “multi-great powers in different parts of the
President William Lai (賴清德) recently attended an event in Taipei marking the end of World War II in Europe, emphasizing in his speech: “Using force to invade another country is an unjust act and will ultimately fail.” In just a few words, he captured the core values of the postwar international order and reminded us again: History is not just for reflection, but serves as a warning for the present. From a broad historical perspective, his statement carries weight. For centuries, international relations operated under the law of the jungle — where the strong dominated and the weak were constrained. That