Saturday, October 17, 2015

Classic polls leave to rest? Researchers have something better – Polish Radio

The research team led by prof. Vladimir Gogołka of the University of Warsaw, several years ago adopted the so-called. Data refining Big Data, based on the forecasts, among others, the results of parliamentary and presidential elections.

– winnowing valuable information from Big Data requires some specialized software tools. Their function is to collect entries – information from the network (doing the work), searching for phrases containing a specific name, eg. A company name that are in the vicinity of words referred to as sentiment. Eg. “Abacki policies is a good economist”. By counting the number of phrases with positive and negative sentiment (in the example of the positive sentiment we are “good”) we obtain the opinion of Abackim, eg. 100 thousand. bad good reviews in 1000 – said Gogołek.

Professor with the team used the method of refining large datasets on the occasion of presidential and parliamentary elections in 2011. It then proved its high reliability. A similar analysis was also performed during the finals of the campaign in this year’s presidential election.

– The parliamentary and presidential elections in 2011 results were provided flawlessly. In the last presidential election meaningful is the percentage difference (just 0.66 per cent.) Between the numbers of positive sentiments regarding each candidate, collected by our tools on the eve of presidential elections in 2015, which was 2.44 per cent., And the real difference that separated Andrzej Duda and Bronislaw Komorowski – 3.10 percent. – Pointed professor.

The expert explained that refining large datasets provides a valuable alternative to quantitative surveys, and by automating processes, its cost in comparison to conventional methods is much smaller.

– classic study based on the analysis, the most commonly categorized answers to questions that are asked specific representative number, hundreds, rarely thousand people. Refining are subjected to while millions of entries. Eg. In recent studies of John Paul II, refining gave up approx. 5 million entries. The reliability of the test classic is the representativeness of the sample, eg. A thousand people in the refining credibility imply before the results – he stressed.

– Compared to traditional research costs of refining are marginal, especially if you have a relatively standardized tools: robots that collect Entries, identification sentiments, sentiments calculation times – noted the professor.

Refining Big Data should not be considered as a research experiment. Gogłek argues that it and similar tools are commercial applications.

– Refining covers a very broad spectrum of possible research m.in .: monitor brand – to identify current threats positive image of the brand, the possibility of collecting sentiments regarding the trading of listed companies. Similar to studies brand is easy, using the refining, monitoring trading organizations, parties and individuals. Identification of threats: crime, the drawbacks of mass products, etc. – He said.

A survey conducted this year by IBM’s Institute for Business Value among executives of global companies showed a strong need for various types of analytics data into the everyday practices of functioning of enterprises and organization.

 

By Hal Varian, Google’s chief economist, Big Data Scientist, researcher or data, will be one of the most sought after professions in IT over the next decade. It is estimated that already by 2020, the network will grow to 45 zetabajtów. Until then, the gap in the labor market in the US will amount to more than 1.5 million vacancies, waiting to fill by specialists from Big Data.

PAP kk

LikeTweet

No comments:

Post a Comment