miércoles, 7 de agosto de 2013

La red desplaza a los profesores de las aulas

How Big Data Is Taking Teachers Out of the Lecturing Business

Schools and universities are embracing technology that tailors content to students' abilities and takes teachers out of the lecturing business. But is it an improvement?


By Seth Fletcher

When Arnecia Hawkins enrolled at Arizona State University last fall, she did not realize she was volunteering as a test subject in an experimental reinvention of American higher education. Yet here she was, near the end of her spring semester, learning math from a machine. In a well-appointed computer lab in Tempe, on Arizona State's desert resort of a campus, she and a sophomore named Jessica were practicing calculating annuities. Through a software dashboard, they could click and scroll among videos, text, quizzes and practice problems at their own pace. As they worked, their answers, along with reams of data on the ways in which they arrived at those answers, were beamed to distant servers. Predictive algorithms developed by a team of data scientists compared their stats with data gathered from tens of thousands of other students, looking for clues as to what Hawkins was learning, what she was struggling with, what she should learn next and how, exactly, she should learn it.
Having a computer for an instructor was a change for Hawkins. “I'm not gonna lie—at first I was really annoyed with it,” she says. The arrangement was a switch for her professor, too. David Heckman, a mathematician, was accustomed to lecturing to the class, but he had to take on the role of a roving mentor, responding to raised hands and coaching students when they got stumped. Soon, though, both began to see some benefits. Hawkins liked the self-pacing, which allowed her to work ahead on her own time, either from her laptop or from the computer lab. For Heckman, the program allowed him to more easily track his students' performance. He could open a dashboard that told him, in granular detail, how each student was doing—not only who was on track and who was not but who was working on any given concept. Heckman says he likes lecturing better, but he seems to be adjusting. One definite perk for instuctors: the software does most of the grading for them.
At the end of the term, Hawkins will have completed the last college math class she will probably ever have to take. She will think back on this data-driven course model—so new and controversial right now—as the “normal” college experience. “Do we even have regular math classes here?” she asks.
Big Data Takes Education
Arizona State's decision to move to computerized learning was born, at least in part, of necessity. With more than 70,000 students, Arizona State is the largest public university in the U.S. Like institutions at every level of American education, it is going through some wrenching changes. The university has lost 50 percent of its state funding over the past five years. Meanwhile enrollment is rising, with alarmingly high numbers of students showing up on campus unprepared to do college-level work. “There is a sea of people we're trying to educate that we've never tried to educate before,” says Al Boggess, director of the Arizona State math department. “The politicians are saying, ‘Educate them. Remediation? Figure it out. And we want them to graduate in four years. And your funding is going down, too.’”
Two years ago Arizona State administrators went looking for a more efficient way to shepherd students through basic general-education requirements—particularly those courses, such as college math, that disproportionately cause students to drop out. A few months after hearing a pitch by Jose Ferreira, the founder and CEO of the New York City adaptive-learning start-up Knewton, Arizona State made a big move. That fall, with little debate or warning, it placed 4,700 students into computerized math courses. Last year some 50 instructors coached 7,600 Arizona State students through three entry-level math courses running on Knewton software. By the fall of 2014 ASU aims to adapt six more courses, adding another 19,000 students a year to the adaptive-learning ranks. (In May, Knewton announced a partnership with Macmillan Education, a sister company to Scientific American.)
Arizona State is one of the earliest, most aggressive adopters of data-driven, personalized learning. Yet educational institutions at all levels are pursuing similar options as a way to cope with rising enrollments, falling budgets and more stringent requirements for student achievement. Public primary and secondary schools in 45 states and the District of Columbia are rushing to implement new, higher standards in English-language arts and mathematics known as the Common Core state standards, and those schools need new instructional materials and tests to make that happen. Around half of those tests will be online and adaptive, meaning that a computer will tailor questions to each student's ability and calculate each student's score [see “Why We Need High-Speed Schools,” on page 69]. School systems are experimenting with a range of other adaptive programs, from math and reading lessons for elementary school students to “quizzing engines” that help high school students prepare for Advanced Placement exams. The technology is also catching on overseas. The 2015 edition of the Organization for Economic Co-operation and Development's Program for International Student Assessment (PISA) test, which is given to 15-year-olds (in more than 70 nations and economies so far) every three years, will include adaptive components to evaluate hard-to-measure skills such as collaborative problem solving.
Proponents of adaptive learning say that technology has finally made it possible to deliver individualized instruction to every student at an affordable cost—to discard the factory model that has dominated Western education for the past two centuries. Critics say it is data-driven learning, not traditional learning, that threatens to turn schools into factories. They see this increasing digitization as yet another unnecessary sellout to for-profit companies that push their products on teachers and students in the name of “reform.” The supposedly advanced tasks that computers can now barely pull off—diagnosing a student's strengths and weaknesses and adjusting materials and approaches to suit individual learners—are things human teachers have been doing well for hundreds of years. Instead of delegating these tasks to computers, opponents say, we should be spending more on training, hiring and retaining good teachers.
And while adaptive-learning companies claim to have nothing but the future of America's children in mind, there is no denying the potential for profit. Dozens of them are rushing to get in on the burgeoning market for instructional technology, which is now a multibillion-dollar industry [see box at left]. As much as 20 percent of instructional content in K–12 schools is already delivered digitally, says Adam Newman, a founding partner of the market-analysis firm Education Growth Advisors. Although adaptive-learning software makes up only a small slice of the digital-instruction pie—around $50 million for the K–12 market—it could grow quickly. Newman says the concept of adaptivity is already very much in the water in K–12 schools. “In K–12, the focus has been on differentiating instruction for years,” he says. “Differentiating instruction, even without technology, is really a form of adaptation.”
Higher-education administrators are warming up to adaptivity, too. In a recent Inside Higher Ed/Gallup poll, 66 percent of college presidents said they found adaptive-learning and testing technologies promising. The Bill & Melinda Gates Foundation has launched the Adaptive Learning Market Acceleration Program, which will issue 10 $100,000 grants to U.S. colleges and universities to develop adaptive courses that enroll at least 500 students over three semesters. “In the long term—20 years out—I would expect virtually every course to have an adaptive component of some kind,” says Peter Stokes, an expert on digital education at Northeastern University. That will be a good thing, he says—an opportunity to apply empirical study and cognitive science to education in a way that has never been done. In higher education in particular, “very, very, very few instructors have a formal education in how to teach,” he says. “We do things, and we think they work. But when you start doing scientific measurement, you realize that some of our ways of doing things have no empirical basis.”
The Science of Adaptivity
In general, “adaptive” refers to a computerized-learning interface that constantly assesses a student's thinking habits and automatically customizes material for him or her. Not surprisingly, though, competitors argue ferociously about who can claim the title of true adaptivity. Some say that a test that does nothing more than choose your next question based on whether you get the item in front of you correct—a test that steers itself according to the logic of binary branching—does not, in 2013, count as fully adaptive. In this view, adaptivity requires the creation of a psychometric profile of each user, plus the continuous adjustment of the experience based on that person's progress.
To make this happen, adaptive-software makers must first map the connections among every concept in a piece of learning material. Once that is done, every time a student watches a video, reads an explanation, solves a practice problem or takes a quiz, data on the student's performance, the effectiveness of the content, and more flow to a server. Then the algorithms take over, comparing that student with thousands or even millions of others. Patterns should emerge. It could turn out that a particular student is struggling with the same concept as students who share a specific psychometric profile. The software will know what works well for that type of student and will adjust the material accordingly. With billions of data points from millions of students and given enough processing power and experience, these algorithms should be able to do all kinds of prognostication, down to telling you that you will learn exponents best between 9:42 and 10:03 a.m.
They should also be able to predict the best way to get you to remember the material you are learning. Ulrik Juul Christensen, CEO of Area9, the developer of the data-analysis software underpinning McGraw-Hill's adaptive LearnSmart products, emphasizes his company's use of the concept of memory decay. More than two million students currently use LearnSmart's adaptive software to study dozens of topics, either on their own or as part of a course. Research has shown that those students (all of us, really) remember a new word or fact best when they learn it and then relearn it when they are just on the cusp of forgetting it. Area9's instructional software uses algorithms to predict each user's unique memory-decay curve so that it can remind a student of something learned last week at the moment it is about to slip out of his or her brain forever.
Few human instructors can claim that sort of prescience. Nevertheless, Christensen dismisses the idea that computers could ever replace teachers. “I don't think we are so stupid that we would let computers take over teaching our kids,” he says.
Backlash
In March, Gerald J. Conti, a social studies teacher at Westhill High School in Syracuse, N.Y., posted a scathing retirement letter to his Facebook page that quickly became a viral sensation. “In their pursuit of Federal tax dollars,” he wrote, “our legislators have failed us by selling children out to private industries such as Pearson Education,” the educational-publishing giant, which has partnered with Knewton to develop products. “My profession is being demeaned by a pervasive atmosphere of distrust, dictating that teachers cannot be permitted to develop and administer their own quizzes and tests (now titled as generic ‘assessments’) or grade their own students' examinations.” Conti sees big data leading not to personalized learning for all but to an educational monoculture: “STEM [science, technology, engineering and mathematics] rules the day, and ‘data driven’ education seeks only conformity, standardization, testing and a zombie-like adherence to the shallow and generic Common Core.”
Conti's letter is only one example of the backlash building against tech-oriented, testing-focused education reform. In January teachers at Garfield High School in Seattle voted to boycott the Measures of Academic Progress (MAP) test, administered in school districts across the country to assess student performance. After tangling with their district's superintendent and school board, the teachers continued the boycott, which soon spread to other Seattle schools. Educators in Chicago and elsewhere held protests to show solidarity. In mid-May it was announced that Seattle high schools would be allowed to opt out of MAP, as long as they replaced it with some other evaluation.
It would be easy for proponents of data-driven learning to counter these protests if they could definitely prove that their methods work better than the status quo. But they cannot do that, at least not yet. Empirical evidence about effectiveness is, as Darrell M. West, an adaptive-learning proponent and founder of the Brookings Institution's Center for Technology Innovation, has written, “preliminary and impressionistic.” Any accurate evaluation of adaptive-learning technology would have to isolate and account for all variables: increases or decreases in a class's size; whether the classroom was “flipped” (meaning homework was done in class and lectures were delivered via video on the students' own time); whether the material was delivered via video, text or game; and so on. Arizona State says 78 percent of students taking the Knewton-ized developmental math course passed, up from 56 percent before. Yet it is always possible that more students are passing not because of technology but because of a change in policy: the university now lets students retake developmental math or stretch it over two semesters without paying tuition twice.
Even if proponents of adaptive technology prove that it works wonderfully, they will still have to contend with privacy concerns. It turns out that plenty of people find pervasive psychometric-data gathering unnerving. Witness the fury that greeted inBloom earlier this year. InBloom essentially offers off-site digital storage for student data—names, addresses, phone numbers, attendance, test scores, health records—formatted in a way that enables third-party education applications to use it. When inBloom was launched in February, the company announced partnerships with school districts in nine states, and parents were outraged. Fears of a “national database” of student information spread. Critics said that school districts, through inBloom, were giving their children's confidential data away to companies who sought to profit by proposing a solution to a problem that does not exist. Since then, all but three of those nine states have backed out.
This might all seem like overreaction, but to be fair, adaptive-education proponents already talk about a student's data-generated profile following them throughout their educational career and even beyond. Last fall the education-reform campaign Digital Learning Now released a paper arguing for the creation of “data backpacks” for pre-K–12 students—electronic transcripts that kids would carry with them from grade to grade so that they will show up on the first day of school with “data about their learning preferences, motivations, personal accomplishments, and an expanded record of their achievement over time.” Once it comes time to apply for college or look for a job, why not use the scores stored in their data backpacks as credentials? Something similar is already happening in Japan, where it is common for managers who have studied English with the adaptive-learning software iKnow to list their iKnow scores on their resumes.
This Is Not a Test
It is far from clear whether concerned parents and scorned instructors are enough to stop the march of big data on education. “The reality is that it's going to be done,” says Eva Baker, director of the Center for the Study of Evaluation at the University of California, Los Angeles. “It's not going to be a little part. It's going to be a big part. And it's going to be put in place partly because it's going to be less expensive than doing professional development.”
That does not mean teachers are going away. Nor does it mean that schools will become increasingly test-obsessed. It could mean the opposite. Sufficiently advanced testing is indistinguishable from instruction. In a fully adaptive classroom, students will be continually assessed, with every keystroke and mouse click feeding a learner profile. High-stakes exams could eventually disappear, replaced by the calculus of perpetual monitoring.
Long before that happens, generational turnover could make these computerized methods of instruction and testing, so foreign now, unremarkable, as they are for Arizona State's Hawkins and her classmates. Teachers could come around, too. Arizona State's executive vice provost Phil Regier believes they will, at least: “I think a good majority of the instructors would say this was a good move. And by the way, in three years 80 percent of them aren't going to know anything else.”
Take an adaptive quiz on state capitals at ScientificAmerican.com/aug2013/learn-smart

Visualización: Estructuras de comunidad radiales

Pinwheel Layout to Highlight Community Structure

Bernie Hogan
Research Fellow, Oxford Internet Institute

Pinwheel Layout to Highlight Community Structure

Caption
The intuitive appeal of force-directed layouts is due in part to their ability to represent underlying community structures. Such diagrams show dense pockets of nodes with bridges connecting across clusters. Yet, it is possible to start, rather than end, with community structure. This is a “pinwheel” diagram using the author’s Facebook personal network (captured July 15, 2009). Nodes represent the author’s friends and links represent friendships among them. The author is not shown. Each ‘wing’ radiating outwards is a partition using a greedy community detection algorithm (Wakita and Tsurumi, 2007). Wings are manually labelled. Node ordering within each wing is based on degree. Node color and size is also based on degree. Nodes position is based on a polar coordinate system: each node is on an equal angle of n/360º with a radius being a log-scaled measure of betweenness. Higher values are closer to the center indicating a sort of cross-partition ‘gravity’.
This layout has several notable features:

                - The angle of each wing is proportionate to its share of the network. Thus 25 percent of nodes go from 0 to 90º.
                - Partitions are distinguished by their position rather than a node’s color or shape.
                - The tail indicates the periphery of each partition. A wing with many tail nodes indicates many people who are only tied to other group members.
                - Edges crossing the center show between-partition connections. Since nodes are sorted by degree it is easy to see if edges originate from the most highly connected nodes or the entire partition.

                This visualization is oriented towards well-connected modular networks (meaning they are easily partitioned into distinct communities). Facebook egocentered networks often have these properties, whereby each partition represents a life course stage or social context and close friends link between partitions.
                In this network it is easy to see a strong series of linkages between high school and university as well as high school and family. There are many ties between the current co-workers and professional colleagues, and neither connects substantially to high school. While just as populous, the professional partition is far less dense than the high school partition.

Self-Commentary
                The data was captured using a custom-built publicly available Facebook application. This application employs the Facebook API to query for a user’s friends and the connections between these friends. The clustering and layout was done using NodeXL, a network analysis add-on for Excel 2007.
                To create this diagram, I “hacked” many of the features of NodeXL. For example, to layout nodes within each partition according to degree I had to first convert the cluster names to cluster numbers. Since degree has a maximal value of n-1, I multiplied each node’s cluster number by n and then added degree thereby ensuring no overlaps. The polar coordinate system does not pay attention to the layout order, so I first laid the nodes out using a circle layout and converted the X and Y coordinates to Radians. The betweenness values only span half of the radius rather than the full radius because otherwise links between adjacent communities would look messy.
                The viewer may notice that I used degree for three metrics: within-partition ordering, size and color. The combination of all three gives the nodes its wing-like shape and gradient. Regardless of the metric used for within-cluster ordering, I recommend also using it for color and shape as well.
                NodeXL is very flexible but it still has many limitations. For example, it currently only includes one community detection algorithm (although it is simple enough to paste in other partitions done through other packages). This particular algorithm misclassified a handful of nodes leading to unnecessary edge-crossings.

PEER REVIEW COMMENT No. 1
In this representation of Facebook, partitions are denoted by position rather than color – which is a clever way to layer more information into the figure. The groups breakdown along foci, or social contexts -high school, camp, college, work etc, and the pinwheel figure nicely illustrates the different foci in ego's network.  However, one wishes the highly interconnected foci were positioned closer
together along the wheel in some way that would minimize the edge-crossing in the middle.  I wonder if the duplication of degree in size, color and position is necessary, or if one could usefully integrate other information through color (say ratio of ties within and between foci), to increase the information content of the image.

PEER REVIEW COMMENT No. 2
The pinwheel layout highlights the organizationally focused nature of friendship ties in the author’s ego network. The diagram captures the relationships that span organizational settings, and more importantly, it does a very good job of showing how densely (or sparsely) the settings are connected with each other. This effect is aided by having the set of ties related to each setting be sized proportional to its contribution to the ego network. The visualization may be improved by imposing some sort of order on where the settings themselves are placed, so that they are chronological or so that settings that share more ties are next to each other.

PEER REVIEW COMMENT No. 3
This submission successfully rebels against the convention force-directed layout, reengineering the circular layout in an information-rich way.  The degree ordering is effective in allowing one to get a sense of how different partitions connect to each other, both in terms of number and distribution of cross-ties.  I'd love to see if this layout would work semi-circle (which might permit degree ordering to be arranged in a roughly up-down fashion while still preserving its radial merits).

lunes, 5 de agosto de 2013

Visualización: Tópicos compartidos por los noticieros estadounidenses

U.S. News Outlets - Network of Shared Topics

Katherine Ognyanova
Doctoral Candidate at University of Southern California, Annenberg

News Ties: A Study of Media-Level Agenda-Setting


Self-Commentary
The research presented here proposes an exploration of inter-media agenda-setting. One of the big questions in the field is whether proliferation of channels and information overload have fragmented news content, rendering the idea of a consistent mainstream media agenda obsolete. The study employs network analysis methods to test this proposition. News outlets are seen as a set of interconnected nodes with ties based on overlap of issue coverage. Two news sources are linked if there is an above average similarity in their selection of topics over a period of time. The research demonstrates that mainstream media outlets not only have overlapping agendas at a given point in time, but also exhibit patterns of shared coverage that persist over time. A QAP analysis comparing networks of shared topics compiled for January and December of 2007 shows a strong and significant correlation (r=0.87, p<0.001) between the two. Furthermore, both networks are dense, well-connected and have a relatively high centralization index - something that goes against the idea of fragmentation and disintegration of media agenda.

The study goes on to explore the set of properties that are likely to influence the overlap of media agendas. Factors taken into consideration include format (newspaper, radio, network TV, cable TV, Internet), common ownership, and volume of the outlet production. An ERG model including parameters for those factors as well as basic structural signatures is fitted to the December 2007 news outlet network. Results of the estimation indicate that both production volume and common ownership affect the topic overlap of news outlets. In order to investigate the patterns of shared coverage within and across media sectors, a blockmodel analysis is performed (results presented on the figure above). Network TV, Internet and radio outlets are likely to produce coverage similar to that of other media of the same type. This does not hold true for newspapers and cable TV. The Network TV outlets have dense ties with all of the other four sectors – a finding consistent with previous research on the influence of television news.

PEER REVIEW COMMENT No. 1
This submission shows the topic similarity between news outlets, where a tie exists if two organizations cover a similar array of topics.  The pairing of the block-model results and the detail graph is very helpful, as the detailed layout is so dense as to make identifying positions challenging.    I wonder if the image would be improved by forcing the nodes into a positioning more similar to the block layout or somehow incorporating a weighting that differentiates within and between-block ties? 

PEER REVIEW COMMENT No. 2
The visualization uses a block-modeling approach to show that despite a proliferation of media outlets, there is much convergence on news topics. While the block model image portrays a general sense of the interconnectedness of the media system, the network image itself is somewhat swamped by density of ties, with little information aside from the size and centrality of the media outlet evident from the detailed image.   The rendering itself is nicely done, with a color scheme that is easy to view and differentiate. 

PEER REVIEW COMMENT No. 3
This visualization has excellent theoretical framing, providing an interesting question, and an easily understood visual answer.  The blockmodel and adjacency plots compliment each other well, and give us both a sense of connectedness and structure.  The use of loops in the blockmodel is especially effective at giving us a sense of internal connectedness compared to connectedness between.  I'd be interested to see a hybrid of the blockmodel and adjacency views.

domingo, 4 de agosto de 2013

Brandes y Pich (2013): Visualización exploratoria de patrones de citas en investigación de redes sociales

Visualización: Enlaces dentro y entre 5 posiciones equivalentes en el comercio de productos agrícolas

Ties within and between the 6 regularly equivalent positions in the trade of agricultural products

Carl Nordlund
Lund University

Ties Within and Between the 6 Regularly Equivalent Positions in the Trade of Agricultural Products
 

Figure 8.11: Ties within and between the 6 regularly equivalent positions (applying criteria fulfillment formula 1 with a relative cutoff value of ~0.0102 (1/98)).

Self-Commentary
The diagram above is a structural map of international flows of agricultural hectares (aka ‘ecological footprints’) between 1995-99. Countries are categorized according to their sharing of similar regular roles in the trade of agricultural products. Positions are determined using multi-dimensional scaling on the matrix containing REGE coefficients, whereas arrows indicate regular ties depicted using a criteria-fulfillment heuristic developed for datasets with huge value spans (Nordlund 2007).



Ties within and Between 8 Regularly Equivalent Positions
Figure 9.8: Ties within and between the 8 regularly equivalent positions: forestry commodity biomass flows (applying criteria fulfillment formula 1 with a relative cutoff value of ~0.0112 (1/89)).

                Similar to the first diagram, the above second figure is a structural map of international flows of timber biomass during the 1995-99 period. Determining role-equivalence through the REGE algorithm, the resulting structural map depicts a structure where a dual core controls the network of forestry product trade.
                Both diagrams are taken from Ecography; International trade, network analysis, and ecological conceptualizations of unequal exchange – my PhD thesis in human ecology (Lund university, Sweden), to be published in August-September 2009.

PEER REVIEW COMMENT No. 1
The visualization above depicts the international flows of agricultural and forestry products. The vast web of trade is simplified by placing countries in regular equivalent roles and mapping the relations across the roles.  It is clear from the second graph that the forestry trade network exhibits a dual core structure, where two sets of regularly equivalent countries are at the core of the network. The regular equivalence graph nicely simplifies a very complicated network.  The image itself is difficult to use in stand-alone form, since too much is assumed of the viewer. It is not clear, for example, what the "criteria fulfillment" actually is and why it is important, so some simple explanatory work there might be helpful.

PEER REVIEW COMMENT No. 2
This visualization is a depiction of the role structure of an international trade network. While this image provides an excellent concentration of a large amount of information, it is lacking a clear interpretation.   As a model of the world system, it would help if there were a simple way to compare the two contents of trade represented in the two figures. I find myself comparing the country codes of the two cores, to see who is left out in each.

PEER REVIEW COMMENT No. 3
This submission provides a great structural view of the data, without obscuring the identities of individual nodes.  I wonder if the node identities can be arrayed to provide additional information, without detracting from the structural clarity provided in an REGE-based diagram.

Journal of Social Structure

sábado, 3 de agosto de 2013

Argentina el 25% de las cuentas twittean 2% menos que la media mundial

Los países más activos en Twitter


De media, tan sólo una de cada cuatro cuentas en Twitter mantiene actividad de forma regular. Esta cifra, no obstante, varía notablemente en función de determinados países.

Según Semiocast, España es uno de los 3 países más activos en la red social del mundo. El 29% de las cuentas de este país twittean con frecuencia, un 2% por encima de la media mundial.

Los Países Bajos y Japón nos adelantan con un 33% y un 30% de cuentas con actividad frecuente, respectivamente. Estados Unidos, por el contrario, ha quedado relegado al cuarto puesto, con un 28% de actividad regular.

Alemania, Filipinas y la India se encuentran a la cola de este particular ranking, en donde la diferencia con respecto a la media mundial puede llegar a ser de hasta un -8%.

Ranking: