Making big data manageable

Author Topic: Making big data manageable  (Read 1242 times)

Offline Rubaida Easmin

  • Full Member
  • ***
  • Posts: 150
  • Test
    • View Profile
Making big data manageable
« on: April 20, 2017, 10:20:10 PM »
One way to handle big data is to shrink it. If you can identify a small subset of your data set that preserves its salient mathematical relationships, you may be able to perform useful analyses on it that would be prohibitively time consuming on the full set.
The methods for creating such “coresets” vary according to application, however. Last week, at the Annual Conference on Neural Information Processing Systems, researchers from MIT’s Computer Science and Artificial Intelligence Laboratory and the University of Haifa in Israel presented a new coreset-generation technique that’s tailored to a whole family of data analysis tools with applications in natural-language processing, computer vision, signal processing, recommendation systems, weather prediction, finance, and neuroscience, among many others.
“These are all very general algorithms that are used in so many applications,” says Daniela Rus, the Andrew and Erna Viterbi Professor of Electrical Engineering and Computer Science at MIT and senior author on the new paper. “They’re fundamental to so many problems. By figuring out the coreset for a huge matrix for one of these tools, you can enable computations that at the moment are simply not possible.”

http://news.mit.edu/2016/making-big-data-manageable-1214

Offline Nazia Nishat

  • Full Member
  • ***
  • Posts: 132
  • Test
    • View Profile
Re: Making big data manageable
« Reply #1 on: April 20, 2017, 10:34:55 PM »
The cluster that contains “dress,” “brides,” “bridesmaids,” and “wedding,” for instance, appears to denote the topic of weddings; the cluster that contains “gun,” “fired,” “jammed,” “pistol,” and “shootings” appears to designate the topic of shootings.
Impressive post...