Aleksandr Kogan knows all too well how a tie-up between academia and the tech world can go very, very wrong.
He is the University of Cambridge neuroscientist who earlier this year achieved international infamy for passing on Facebook profile data from tens of millions of users to the parent company of Cambridge Analytica, the now-defunct political consultancy accused of using this information to target potential Donald Trump voters in the 2016 US presidential election (something the firm has denied).
Now living in New York and working on a new online survey tool, Kogan acknowledges that the bad publicity has in effect ended his academic career. 鈥淚 completely missed how people were going to react,鈥 he says.
For some, this might sound like just deserts. Kogan used a Facebook app to harvest profile data from not only those who installed the app but their unwitting Facebook friends, too (something no longer possible after Facebook changed its rules in 2014). Some have suggested that his colleagues thought that what he was doing was unethical; Kogan had an application to use the data for academic purposes rejected in 2015 over concerns about consent.
探花视频
As Kogan tells it, he was naive 鈥 but not greedy or ethically lax. He says he did not personally gain financially from his Cambridge Analytica deal, but simply wanted more funding to help gather a juicy research database from the world鈥檚 biggest social network. He had 鈥渘o inkling鈥 anyone would be upset.
鈥淎ll this seems unbelievable and silly now, but from that vantage point it seemed sensible,鈥 he says. 鈥淚鈥檓 a sceptical scientist [but] not a sceptical person.鈥
探花视频
Neither his manager nor colleagues raised ethical objections to the tie-up, Kogan insists (a Cambridge spokesman said he could not comment as this would constitute personal information). 鈥淭he university is very encouraging of its faculty members to go and do entrepreneurial activity,鈥 partly as a way to hit impact targets in the research excellence framework, he adds.
Yet Kogan鈥檚 鈥渆ntrepreneurial activity鈥 culminated in denunciation of the university on the biggest stage possible. Facebook founder Mark Zuckerberg, called before the US Congress after the scandal blew up in April, asked 鈥渨hether there is something bad going on at Cambridge University overall that will require a stronger action from us鈥.
Earlier this month, the UK鈥檚 data protection watchdog, the Information Commissioner鈥檚 Office, revealed that it is investigating whether Kogan has committed a criminal offence. It announced that it is to audit the Cambridge Psychometrics Centre, where Kogan worked, for compliance with the Data Protection Act. The ICO is also to carry out, with Universities UK, a broader review of academics鈥 use of personal data, in both their research and commercial capacities. And the office has fined Facebook the maximum possible 拢500,000 over its part in the Cambridge Analytica scandal.
In response to Zuckerberg鈥檚 question, Cambridge that it had worked for years on publicly available research that used Facebook data, including studies co-authored by Facebook employees.聽Kogan's venture聽is only one聽example - and one that went most specularly wrong - of a tie-up between聽academia and big tech. Earlier this year, for example, France鈥檚 脡cole Polytechnique announced a new chair in artificial intelligence 鈥 funded by Google.
But the Cambridge Analytica scandal is just one instance of the 鈥 鈥 that many observers consider to be under way against the likes of Google, Facebook, Uber and Amazon. Silicon Valley鈥檚 finest stand accused of a litany of failings, including providing a platform for Russian interference in foreign elections, monopolistic behaviour, workforce exploitation 鈥 and simply making us feel miserable through the relentless interpersonal comparisons facilitated by social media. 鈥 鈥 ran one headline on Buzzfeed late last year.

So should reputation-conscious universities reassess how they work with Big Tech?
James Williams has worked on both sides of the fence. A former Google advertising strategist, he is now a doctoral candidate at the University of Oxford鈥檚 Oxford Internet Institute. His new book, Stand Out of Our Light, warns that Silicon Valley鈥檚 tools of distraction risk undermining our personal and collective will and freedom.
Williams thinks that researchers and universities that are funded by tech firms, or dependent on their data, are yet to apply the same 鈥渟ensitivity鈥 over conflicts of interest that is normal in, say, pharmaceutical research. 鈥淭here is a sexiness to tech companies that鈥檚 obscured these questions of the power dynamics,鈥 he says.
探花视频
One of the earliest examples of the relationship turning sour came in 2014, with the publication of a , authored by researchers from Facebook and Cornell University, that involved manipulating the moods of more than 600,000 Facebook users by exposing them to positive or negative emotions. The study, 鈥淓xperimental Evidence of Massive-Scale Emotional Contagion Through Social Networks鈥, published in PNAS, constituted a huge experiment on subjects whose consent was not sought, and triggered a major against Facebook, the university and the journal.
Kogan, who collaborated with the social network until 2015, says that 鈥淔acebook has data that can answer any question I鈥檓 interested in鈥. But he recalls that the social network became 鈥渋ncreasingly conservative鈥 about working on academic papers in the wake of the reaction to the PNAS study. 鈥淭hat paper gave [them] so much negative attention that they clamped down hard on anything being published,鈥 he says.
Now, following the Cambridge Analytica scandal, Facebook is introducing further restrictions. Previously, academics could gather anonymous data about user behaviour, but Facebook is 鈥渟hutting [that] completely down鈥, according to Anja Bechmann, an associate professor at Aarhus University in Denmark, who studies social media and artificial intelligence and is one of dozens of academics to sign a earlier this year warning that the changes will stymie academic research.
鈥淚f we want data, we have to work with [Facebook] directly鈥, Bechmann says. The fear is that 鈥渙nly the lucky few鈥 will be permitted to do so: namely, the most prominent scholars from the most famous universities 鈥 and primarily those based in the US. Facebook will in effect be able to 鈥渃hoose the assigned research and research question and team鈥, she warns.
This is particularly unfortunate since, as social life has moved online, social media offers a much richer dataset than things such as traditional censuses for social science research, says Bechmann 鈥 who has compiled a lengthy list of publications that she says could not have existed without access to social media 鈥渁pplication programming interfaces鈥 (APIs), which third parties use to glean data from these sites.
鈥淚t鈥檚 not good for democracy or our understanding of society that [the public] don鈥檛 have access to research on [social media] data,鈥 she says.
Facebook did not respond to questions from THE.
The issue of access to data goes wider than social media companies, however. For instance, the race is on in Silicon Valley to create safe, reliable self-driving cars 鈥 to which end, companies like Google and Uber have amassed mountains of street-level images and scans from roving test cars in order to teach self-learning artificial intelligence how to deal with every conceivable situation on the road.
This data 鈥 a potential treasure trove 鈥 鈥渋s being held on to very carefully鈥 by the companies that collect it, says Andrew Moore, dean of Carnegie Mellon University鈥檚 School of Computer Science, despite a recommendation from the Obama administration that it should be made publicly available.
But in other data-heavy areas, researchers don鈥檛 need a relationship with a tech company, Moore points out 鈥 medical data, say, comes from teaching hospitals. And nine in 10 Carnegie Mellon academics are 鈥渜uite satisfied鈥 with the access they have to big data, he says: 鈥淲e get approached very frequently by companies who want us to help them with large amounts of data, as opposed to us going out begging for data and the companies saying no.鈥
But Carnegie Mellon has suffered its own tech-related headaches. In 2015, Uber left Moore鈥檚 department 鈥渟crambling to recover鈥 after tempting 40 academics and technicians away with huge salary bumps to form a lab in Pittsburg, The Wall Street Journal . There was a 鈥渢ough period of three weeks when we were trying to figure out how we are going to move forward with our research鈥, Moore said at the time. Defections have continued to occur: earlier this year, for example, the department lost Manuela Veloso, head of machine learning technology, to financial services company JPMorgan Chase.

Stemming this tide of researchers to the tech world has become a big issue for universities, particularly in hyped, lucrative areas like artificial intelligence. It was a key concern of a recent national 础滨听谤别辫辞谤迟 released in France, which recommended 鈥 highly optimistically, as one of the authors admitted 鈥 doubling the salaries of graduate students in this area to stop them leaving.
Moore, who himself worked at Google for eight years, describes his job as 鈥渓ike managing a sports team. You鈥檙e going to be recruiting many folks, but you don鈥檛 expect them to stick around forever鈥. His main strategy for recruitment and retention is to appeal to researchers鈥 idealism. It remains easier, typically, to change society for the better in an academic rather than a corporate position, he says. For instance, computer science researchers at Carnegie Mellon have created an that scans online adverts to detect and identify sex traffickers in the US, facilitating 鈥渁lmost daily arrests鈥. That is 鈥渋ncredibly rewarding鈥 for the academics behind it, Moore says.
And while corporate researchers may not have to write government grant proposals, nor do they have access to unlimited resources. 鈥淭he sadness is that you see them getting really excited about getting hold of a single intern for three months in the summer 鈥 whereas professors get to work with five to 10 graduate students,鈥 Moore notes.
Still, in many cases, working for a big tech company can be the best way to get a new tool into the public domain, he admits. 鈥淚t鈥檚 not just the data, but the access to the channels to take an idea and get it released to millions of users. That is very exciting.鈥
Then there is the money question. Last November, The Guardian on fears among AI researchers that 鈥渢he cr猫me de la cr猫me of academia has been bought鈥 by Silicon Valley. In one case, for instance, Apple convinced a PhD student at Imperial College London to drop their studies for a six-figure salary.
Nor is that cr猫me de la cr猫me confined to technological fields. Tech firms have also taken to hiring university economists. According to Susan Athey, professor of the economics of technology at Stanford University, this is not only because they want to better understand the complexities of online markets, but also because they feel a need to counter the looming threat of anti-monopoly regulation. 鈥淚n-house economists can directly inform regulators and also help outside economic experts learn about the institutional facts, access data and become informed鈥, Athey has . 鈥淓very week, I am contacted to help fill a position, or I hear about a new hire by firms like Airbnb, Netflix, [music streaming service] Pandora or Uber.鈥
But it remains computer scientists that tech firms most crave. According to Moore, researchers with experience of building autonomous systems 鈥 such as robots that can work underground 鈥 number in the 鈥渇ew hundreds鈥 globally, and are consequently like 鈥済old dust鈥 to companies. Moving to a tech company nets such people a compensation package three to five times what they could earn at a university.
The result is that in a faculty of around 200, Moore loses 10-15 people a year to industry, with only around five coming the other way. This has required him to hire about 50 new academics in the past three years. His point to his recruits is that they should see the revolving door as a plus: 鈥淵ou can do these round trips,鈥 he tells them.
探花视频
But is a revolving door really a healthy state of affairs? in The New York Times last year, the data scientist Cathy O鈥橬eil warned that one consequence is that 鈥減rofessors working in computer science and robotics departments 鈥 or law schools 鈥 often find themselves in situations in which positing any sceptical message about technology could present a professional conflict of interest鈥. For this reason, academia is 鈥渁sleep at the wheel鈥 when it comes to warning lawmakers about tech鈥檚 downsides, she added.
Her article attracted strong rebuttals on social media, particularly from academics in the humanities and social sciences, who pointed to their often robust criticism of tech firms. But Moore admits that the revolving door does indeed create 鈥渟omewhat of a conflict of interest鈥.
鈥淚 don鈥檛 think I would ever come out and make statements against a specific company 鈥 unless, of course, I knew it was doing something really bad,鈥 he admits. 鈥淏ut if a company frustrated me in a particular month, or something like that, it does not make good business sense to moan about it publicly because usually it鈥檚 part of a bigger relationship.鈥

One option for academics who want to work for tech firms but also want to keep a foot in the academic world, are joint appointments. These have become increasingly common. Amazon鈥檚 chief economist, Patrick Bajari, is also professor of economics at the University of Washington, for instance. And in 2014, seven academics from Oxford鈥檚 computer science and engineering departments were recruited by DeepMind, a London-based AI company bought by Google in that same year and best known for creating a program capable of beating humans at the board game Go. Three of the academics 鈥 including Royal Society fellow Andrew Zisserman, a computer vision expert 鈥 also remained professors at Oxford. As part of the deal, Google also gave a 鈥渟ignificant seven-figure sum鈥 to their departments.
Bajari did not respond to a request for an interview.聽探花视频聽also contacted the three DeepMind professors for an interview, but a spokeswoman for the comapny instead provided a statement from Murray Shanahan (pictured left), professor of cognitive robotics at Imperial College London, who is also a researcher at the firm. 鈥淭he major incentive for me [in accepting a joint position] was the chance to pursue my research full time without the drain of other academic duties, with access to fabulous resources and in the company of the best like-minded people on the planet,鈥 Shanahan says.
That motivation could be compared to that which attracted many academics to the fabled Bell Labs in the US, the corporate laboratory whose researchers won eight Nobel Prizes between 1937 and 2014. But are DeepMind researchers entirely free to choose what they research? The company鈥檚 spokeswoman says that the company does not 鈥渋nfluence who researchers with dual affiliations collaborate with outside of DeepMind鈥. But she did not answer questions about whether there are any restrictions on what academics with joint affiliations can publish.
For his part, Shanahan admits that before he joined DeepMind, he 鈥渃onsidered鈥he potential loss of freedom and independence I might experience from being part of a big corporation. 鈥媁ould I still be free to say and do what I liked (for example, to speak to the press) to the extent that I was as a full-time academic?鈥 But after a year of working for the company, it is a case of 鈥渟o far, so good鈥.
Moreover, despite being so potentially lucrative for big tech companies, machine learning and artificial intelligence have remained very open fields. Researchers were up in arms earlier this year, for instance, when Nature Publishing Group proposed a closed-access journal to serve the discipline. 鈥淭he general advantages of being open about research in this area outweigh the potential or perceived advantages of being secretive,鈥 says Zoubin Ghahramani, a professor of information engineering at Cambridge and chief scientist at Uber (the increasing demands of the latter role requiring him to relocate from Cambridge to San Francisco in August).
鈥淥f course, Uber is a company, and so we have to be careful with respect to any commercially sensitive information,鈥 he says. 鈥淪o obviously we wouldn鈥檛 want to publish our business practices, which other companies might be very interested in, and we have to be careful about other things like IP and so on. But the norm鈥s in favour of openness.鈥
In the development of self-driving cars, 鈥渧ery little even in this field is about having a particular secret sauce for something鈥, he argues. 鈥淚t鈥檚 not like one research paper is going to make a huge difference鈥 as to whether one company wins the race to build a reliable vehicle.
But some remain troubled by the potential ethical compromises to which joint appointments could expose academics. This is particularly the case in Germany, where such positions remain hotly contested. The country is keen not to be left behind when it comes to artificial intelligence, and has created a 鈥淐yber Valley鈥 near Stuttgart that brings together university researchers and companies. And Martin Stratmann, president of the Max Planck Society, Germany鈥檚 vast basic research network, explains that it has hired heavily from the US to bring in directors for the Institute for Intelligent Systems, founded in 2011, that constitutes the 鈥渘ucleus鈥 of the project.
But he is dead set against allowing his academics to work for more than one master. 鈥淚n the Max Planck Society鈥e do not have joint appointments,鈥 he says. 鈥淲e want to define our own rules. We have our own ethics rules. We have our own ethics councils 鈥 so we decide where to go.鈥
But others dispute that the ethical strains potentially imposed by joint positions with tech firms are uniquely acute. Uber is currently facing questions over an incident earlier this year in which one of its prototype self-driving cars 鈥 carrying a human observer 鈥 hit and killed a woman crossing the road in Arizona. So what would Cambridge鈥檚 Ghahramani do if he felt the company was rushing out a solution before it was safe?
鈥淚f something didn鈥檛 match my ethical standards, I would speak out,鈥 he insists. 鈥淚 think this is the role of whistleblowers in any kind of situation.鈥
DeepMind鈥檚 Shanahan also denies that 鈥渉aving a joint appointment puts me in a different position to any employee of any company or organisation鈥. Moreover, he does not expect to be confronted by any particularly serious ethical conflicts in his current role.
鈥淥ne of the reasons I鈥檓 comfortable working at DeepMind is that there is a strong ethical ethos to the company,鈥 he says. 鈥淪o I don鈥檛 anticipate having to face such a moral dilemma.鈥澛
Campus, Inc: should universities venture into building businesses?
When the London tech start-up Magic Pony was sold for a reported $150 million (拢102 million) to Twitter in June 2016, just 18 months after it was created, City investors sat up.
Admittedly, it wasn鈥檛 quite the jaw-dropping levels of profit enjoyed by Instagram founder Kevin Systrom, who sold up to Facebook for $1 billion in 2012 after a year of business, but it was further evidence that the UK capital was a growing rival to Silicon Valley for machine learning (Magic Pony uses neural networks to enhance images), following the sale of predictive text company SwiftKey to Microsoft for 拢174 million months earlier and DeepMind to Google for 拢400 million in 2014.
While Magic Pony鈥檚 founders, Rob Bishop and Zehan Wang, were graduates of Imperial College London, they did not fit the familiar 鈥渦niversity pals strike it rich鈥 narrative of Google or Facebook. They met at Entrepreneur First, a business incubator based in south London that seeks to bring together the brightest minds to see if they can come up with businesses that will fly.
鈥淚t鈥檚 a pretty unique model,鈥 believes Joe White, the company鈥檚 chief financial officer, who joined in 2016 having sold Moonfruit, the website building company he co-founded just after graduating from the University of Cambridge, for $40 million in 2014.
鈥淲e鈥檙e sometimes conflated with traditional incubators as our output is similar. The difference is that we bring people together pre-team, pre-idea,鈥 White explains. The 100 recruits in each cohort are offered a 拢2,000-a-month stipend for their first three months, as they set up their companies and develop investment pitches. The 20 or so businesses that look the most promising are then given 拢80,000 and a further three months of support 鈥 including mentoring from successful entrepreneurs and introductions to potential investors and customers 鈥 in return for an 8 per cent stake.
Recruits are typically in their mid- to late twenties, and are often PhD students or postdocs. 鈥淭hey have to have an edge,鈥 explains White, and 鈥渁 deep specialist knowledge鈥 is an obvious example of one. A recent team, for instance, paired a graduate of a PhD in black holes with someone from the finance world to create an AI financial adviser.
That company has already attracted 拢1.6 million in venture capital investment. And with 10 London and three Singapore cohorts now complete, Entrepreneur First has developed companies with a total combined value of 拢1.5 billion, which have raised more than 拢300 million from venture capitalists, White says.
Yet the company鈥檚 heavy reliance on university talent raises the question of whether this is the type of thing UK universities and business schools should be doing themselves. Could such a blurring of the university and tech worlds be one way for them to maintain a connection with their most promising early career researchers in computer science, while also tapping into the vast amounts of money to be made in the tech world?
White demurs. 鈥淯niversities, like many industries, get good at doing a certain thing. They are very good at educating and producing world-class research, [but] the venture capital world is at odds with this environment, so having these things operating beside each other is very difficult.鈥
For him, universities should confine their involvement with the tech world to investing some of their endowments in it. He cites Stanford University鈥檚 lucrative investment in Sequoia Capital, early investors in PayPal, Google and WhatsApp 鈥 although he concedes that few UK universities have remotely comparable amounts of capital to play with.
White also cautions against universities鈥 preference for what he calls the 鈥渋ntellectual property bear hug鈥, whereby they 鈥済rab hold of anything that looks promising鈥, taking large ownership stakes in the spin-off company and thereby 鈥渟tifling鈥 its further growth. In a recent example, a nascent tech company struggled to win seed funding because the PhD student who ran it had given up a 50 per cent share to his institution. The investment was secured only after the university reduced its share to 10 per cent.
鈥淸The PhD student] was a researcher, not an entrepreneur, so didn鈥檛 understand the deal at the time,鈥 White says. 鈥淏ut it meant his idea wasn鈥檛 going to get off the ground 鈥 investors won鈥檛 back something if so much has been given away. [For universities], it鈥檚 better to have 10 per cent of something that becomes massive, than 50 per cent of not very much at all.鈥
探花视频
Jack Grove
POSTSCRIPT:
Print headline: How green is the valley?
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?








