tv The Media Show BBC News September 3, 2023 8:30pm-9:01pm BST
8:30 pm
to reshuffle his shadow cabinet tomorrow as mps return to westminster from their summer break. now on bbc news, the media show: ai — destroyer ofjournalism? hello. ai — it is all we seem to hear about these days, but what does it mean for the news business and the way we all find out about what's going on in the world? what sources will ai rely on to deliver trustworthy news? will it putjournalists out of a job? the chances are you've already, perhaps unknowingly, read a news article that wasn't
8:31 pm
entirely written by a human, so what's going on? today, we're dedicating the whole programme to these questions. with me are madhumita murgia, artificial intelligence editor at the financial times, sky news�* science and technology editor, tom clarke, eliz mizon, from independent media cooperative the bristol cable, as well as jackson ryan, science editor at cnet. welcome to you all. and i think we should start with the basics. madhu, if i could bring you in, from the financial times, explain what we mean by aland why, particularly in terms of the role ofjournalism it has, why it's getting so much coverage now. well, so, ai is artificial intelligence and, i mean, supposedly it's a mechanical computer version of human intelligence, or at least that's the hope, right? but today, what we have is, it's basically a powerful statistical system, a computer software, which finds patterns in large amounts of data. but what this means is that it can, you know, find diagnoses
8:32 pm
from pictures of x—rays or it can look through lots of words and help translate them into different languages. and what we're talking about today is generative ai, which is software that can actually create and generate things that include words, images, code, even video. and how widely is it being used in newsrooms, do you think? i mean, what's the financial times doing, for example? so, i think, over the last six months, it would be impossible to ignore it if you were a newsroom with a digital operation that was trying to reach people online. i think you'd have to be aware and, you know, and have to be experimenting with it. most big, large news publishers are doing it — the ft is. we would... we've put out, our editor put out a letter saying we're not going to be publishing any stories that are written by ai, but we will be looking at how it might help journalists do theirjobs better, things like summarising complex documents, like, you know, tax documents or,
8:33 pm
you know, readouts from court cases, things like that, that are difficult for humans to read lots of, very quickly. it could help to sort of pull out trends, and it doesn't mean it might be great at it. we're trying it out. we will continue to experiment, but i would say nothing, nothing that we're putting out into the world for our readers is ai—generated today. and when it comes to concerns around accuracy and bias, just talk us through that. so, the way that generative ai works, text generation, let's look at, you know, writing words. so, something like chatgpt, which we know you ask it a question and it comes out with an answer. the way it works is it's been trained on billions of words that it's taken from the internet. and those could be words from books, from websites, blogs, reddit posts, youtube comments, think anywhere where there's been words written by humans on the internet. if you think about that corpus of data, you can also see that it's not necessarily fact—checked, accurate.
8:34 pm
in terms of bias, it's also pulling a lot of the sort of implicit assumptions, stereotypes and so on, and all of that is kind of pulled into the software to be trained from. and the way it works is by predicting the next word based on all of the words that it's already been able to analyse. so you can see then why it's not going to be 100% accurate, because it's just telling you what it thinks is most likely based on the past, which is usually true, but not always. 0k. and when we talk about al, it's sometimes discussed as an existential threat to humans. i suppose what we'd be talking about here is whether ai and journalism is going to put us all out of a job. tom clarke, you recently did an experiment for sky news, asking if ai could replace your reporting. let's just hear some of it. and here you're watching a report by a visual avatar whose image is based on a real—life colleague. we associate scenes like this with hotter, drier countries than our own. the next task is to use different types of ai and a human volunteer
8:35 pm
to give our reporter a personality. our producer, hannah, lending her face and voice to train an avatar. i have been trained using a four—minute video clip of hannah speaking into camera. it's pretty convincing to me. yeah, that would fool me. 0k, well, there you are. you were impressed. talk us through what you found by doing those series of reports. they were, yeah, they were visually pretty impressive and, weirdly, we watched them get better during the process of even making the report. so, you know, it was the pace at which things are getting better that also really blew us away. the other thing we tried to look at, though, is these natural language models that madhu was just talking about there and how much potential they have for doing journalism and with the help of someone who understands far more about it than i do, we came up with this little wheeze where we basically got two agents powered by gpt—ii to sort of talk to each other. one played the role of an ai reporter, the other of an editor to kind of pitch stories, refine them, pitch them again, then go and find sources for them,
8:36 pm
you know, another prompt to chatgpt to go off and do that to build something up. and we were feeding it news from this thing called a web crawler to give it some sort of awareness of what's going on out there. and do you know what? it was, it was quite impressive, it was quite cool. it could come up with reasonable sort of pitches for stories, and it could certainly do a really quite convincing job of writing an article. were you encouraged by the things it couldn't do? i mean, were there things that it couldn't do that made you think, "oh, i've got a job for a bit longer"? heaps. so, while it was quite good at coming up with pitches, stories that sort of were credible, they weren't particularly great and i think there's reasons for that. we gave it the news, what was out there in the news, and it was coming with stories which were like, take event x happening in the uk, so house prices and interest rates, oh, there must be a connection. it would sort of take two things and pitch a story around that. it would pitch feature ideas basically, but it doesn't know what news is...
8:37 pm
it wasn't breaking news every day. well, it can't. it's an actual language model designed to predict where the next text was based on sort of training data that's a little bit out of date. it doesn't have any awareness of what's going on in the world beyond what we could feed it from google and the sky news website and other sources we gave it. and it also doesn't have a capacity for abstract thought or imagination or the sort of ideas that we need to make sort of news. the other interesting thing was the hallucination thing, the kind of, where it really gets a bit more worrying. it's really convincing. it does a very good job of presenting you text that's quite believable, but it can be really wrong. so, one story it came up with, there was a lorry crash on the m6 that spilt, i think it was 20,000 litres of milk, all over the m6 motorway. it was a news story and it came up with this pitch that scientists had discovered a hidden benefit of spilling milk on motorways actually made road surfaces safer. and i thought, that's...
8:38 pm
i mean, it was really bizarre. i mean, itjust created this idea. it even found a piece of academic research from a university in new zealand that supported this, that didn't exist. it gave me the academics' names and a journal that it had been published in, and i couldn't find any record of it, and said that this discovery had been made sort of overnight after the accident had happened. absolutely untrue, and is a really good example of where you wouldn't want to be letting an aldo anything, approaching the kind of editorial side ofjournalism. i wanted to bring injackson ryan here to talk about transparency. you work for the american tech website cnet, and they've been using al to help write stories. tell us a little bit about that. i think cnet's approach was a little controversial, but we have been using artificial, a generative ai tool to create articles and then those articles were fact—checked by a human
8:39 pm
and then published on our website. this was what was called an experiment at the time. it actually happened very, very early after gpt sort of exploded across the web. and i think we were kind of like one...very early movers on the generative ai movement. we're a tech website, it seems pretty...like something that we would do, but unfortunately a lot of these articles that were generated by the tool were incorrect and they were generated in an area that we know the tool is not very good at generating text for, which is with numbers. even simple things like, what is a credit card? we were getting an al to generate an answer to that. unfortunately, more than half of those articles that we published were incorrect in some way, needed correction. so we have had to kind of change tact a little bit and... we haven't stopped doing it. we put a big pause on it at cnet and this is one of the things that
8:40 pm
i'm personally quite worried and concerned about in the ai world, it's silicon valley mentality — move fast, break things, see what happens. and a piece i recently wrote, you know, i don't want to compare this too much to the atomic bomb, i had just seen oppenheimer when i wrote this piece, but basically, like, for me, in some ways, it's like standing and watching the gadget, trinity be assembled, right? this first test of... ..kind of a world—changing technology and we haven't really grasped the consequences of what deploying that technology in full means. and unfortunately for us at cnet, we did deploy it without really thinking about what it could mean or perhaps even, i guess, what it could do to some of our credibility, and it was a real harsh lesson that we had to learn. in terms of what you have learned, then, you know, are there things that you're now putting in place at cnet? yeah, yeah, definitely. so, we put a pause on articles once
8:41 pm
this, these articles were discovered and essentially we rewrote a whole ai policy for cnet. the policy now basically states that we will not use ai to write entire articles. we will also not use it for photos or images on our website. but what we will do, it's actually even got a funky name, it's called ramp, which means responsible ai machine partner. and basically, this ramp tool is meant to assist us with creating articles. have i used it in any of the reporting i do as a science editor? no, i don't. there's not anything that i can really use it for, especially when i'm talking about breaking news or new studies, but recently, we published about the best broadband provider in tulsa, oklahoma. and for an article like this, there's probably a lot of work that can be reused and that al tool that we're using is trained on our own data rather than the whole web. now, there are some still some questions that have to be asked, and that's why i'm saying we need
8:42 pm
to slow down with a lot of this stuff being pushed out. just because both of you have raised the fact that, you know, ai generative articles that, you know, you're aware of, or in tom's experiment, actually produced things that were factually inaccurate, why, jackson, was it, was it not getting it right? it trained to generate what the next best word is. it's like a really fancy order predict tool on your phone. your phone learns what you're texting all the time to your friends, to your partner, "i love you" _ it knows you're going to say i love you. and if you go to say, "i," to me and we've just met, it will still tell you, "i love you." it's just the way that the models are trained. and they're also, i think we discovered, they're also really lousy experts. they're written to give you an answer to whatever you ask. if it can't come up with a good answer, it'll make one up. that's the worst kind of expert. a real expert would say, "i'm sorry. i don't have the relevant information. "i don't know." but i think we mustn't underplay how good these language models are. they�* re extremely clever at creating text.
8:43 pm
it's hard to know where you might be looking at one of these hallucinations or not. you know, they're very unreliable experts. you have to be pretty careful because they can give you very convincing wrong answers. jackson, in terms of cnet's journalism on the subject, you now carry a declaration, you know, "how we will use artificial intelligence at cnet." i think the guardian does the same, but how much do you think this matters to audiences? i mean, that's the real, that's the million—dollar question. in some ways, i don't think audiences necessarily care that much where the news comes from. like, this experiment that cnet ran originally was not found for three months, because we had a dropdown box that said, "this was generated with the help of an ai," and it was only that someone had scraped google basically and seen that we were publishing it, that it became known. i don't think audiences even care what bylines are in an article half the time. also, i don't even know that audiences read past the headlines all that often. i don't, i don't want to denigrate
8:44 pm
our audiences because, thank you for reading our site, but at the same time, i don't know that it matters too much and that's scary to me. i would much prefer that that wasn't the case, but i feel like it is. but you are trying to be transparent and, madhu murgia from the ft, i was just going to say, is there the same commitment to transparency across the news industry? well, i'd say i disagree. i think that maybe it's true that people don't recognise different bylines always, but i think people expect that there was a person there who went off and did theirjob, which is to fact—check what they wrote and tell you some version of at least, of the facts or the truth, right? i think there is a sense with audiences of breaking some sort of implicit trust that you have, whether you're broadcast or print media. and i think any media organisation that wants to maintain that relationship of trust will have to be transparent going forward, partly because of the problems around hallucinations and inaccuracy, but also because it's a huge shift in how we as a society
8:45 pm
are consuming information. you know, you can't just go from saying humans no longer do thisjob, this isjust all written by a machine and that's just ok with everybody. i don't know that everyone would accept that. well, let's, let's look at this from a local news perspective. i'm very aware eliz mizon has been sitting there very patiently while this has been going on, from the bristol cable. i mean, ai must be tempting, eliz, for local news publishers. i suppose if you can produce news cheaper than with humans. yeah, i mean, ithink for a lot of people, . probably what we do at the cable is we're, we're really trying to do something different with our co—op and a lot of that isjust to do - with the business model. and i think that there's a bit of, maybe a bit of a paradox - in that we don't necessarily - have the resources to be doing lots and lots of research - into, how can we use ai? you know, if we did use ai — we haven't so far — - i think there's a general feeling that we want to see kind - of where the dust settles, if it indeed settles. -
8:46 pm
but we don't have lots of money and lots of resources to - start experimenting i with this kind of thing. and we are really committed to investigative slow news, if you like. i so there's a bit of a paradox - there that, you know, you'd think, oh, it would be really useful for localjournalism, - which has been worst hit by the kind of collapse, i if you like, of the journalism business model, _ particularly print journalism. but then, at the same time, we don't necessarily- have the resources to be picking up new tools and doing new things - and learning new software. so it's kind of half a... six to one, half - a dozen of the other. i mean, i suppose you're bucking the trend, aren't you? because you say you're very much focused on investigative stories and the sort of proper meaty end ofjournalism. for local news more widely, i suppose, which is often those little, local organisations that are owned by bigger organisations that are looking to cut costs, i suppose that's where this might accelerate trends that are already under way in terms of cost cutting and cutting journalists. yeah, definitely. i think... -
8:47 pm
so, there are kind of two things i that i think are really interesting, and certainly the business model. for me is the most important thing. so, the business model of print journalism particularly- has kind of collapsed. the worst hit of that| are the local outlets. and i think that there - is a situation in which this is going to be really useful for some of those outlets. j so, news corp australia, for- example, recently started using ai in order to essentially kind of aggregate information. i so, one of the really good examples that they used was looking - at the cheapest fuel prices in an area, for example. . now, i wouldn't necessarily call that reporting. - i think that that's really useful and i think they even called it| service information, - providing service information. i think they also made it clear that it was overseen by humans, so they weren't just letting the ai tool going off, go off and print stuff. exactly. so i think that's the ways - in which it can be really useful. but i think what is a problem and what is likely to become | a problem is that the collapse of the business model- is simply going to continue
8:48 pm
if we start thinking, - oh, well, ai canjust do human - beings' jobs and we can sack more people because now we can get chatgpt to write all _ of our articles for us. i think that really misunderstands, particularly with investigative - journalism, how much emotion - there is, how much empathy there is. i mean, one of the examples that i really like to use is it will be - l really useful for an al to take i the minutes of a council meeting and to write that up into an article and then we can fact check it, - but an ai is never going to be able to convince some whistle—blowersj at the council to tell - you what was never minuted in the first place. madhu, injournalism, if we think about the upcoming news cycle, the us election, for example, a general election coming in the uk, we worry about disinformation and deepfakes, but do we need to be concerned about the impact of ai—produced disinformation and fakery that they might have on those stories, for example? definitely. i think this is probably the kind of near—term hot thing that
8:49 pm
everybody is worried about. and we've already seen examples of, you know, politicalfake news, misinformation and disinformation being generated by ai tools. i reported a while ago on... this happened in venezuela, where they had ai—generated news readers reading out government propaganda. much of it wasn't true and it was being generated using a technology based here in the uk. so this stuff goes global really quickly and, you know, there's been hundreds of examples over the years and particularly recently, dozens where we've seen how even images can be manipulated — pictures of borisjohnson being arrested, there was a fake image of trump hugging anthony fauci. so this can be deployed and employed as political tactics. so, until there's some kind of law that forbids it and a way for people to tell real from ai—generated, the flood of it is just
8:50 pm
going to make it harderfor us to tell the difference. tom clarke from sky news? yeah, i wasjust going to add, and we can't underestimate the power of the tools. the same way that they could benefit local news by giving you sort of hyper—local, targeted information, we found it was very easy to get gpt—4 through a little, few prompts to generate emails and send those emails automatically. we did that with sort of off—the—shelf stuff — very unsophisticated. you could, you could have a targeted political ad campaign posting social media on a hyper—local level. you could get inside particular electoral areas with particular memes or messages or whatever in an extremely powerful way. and i think while we injournalism, we're here, we're sort of discussing how it might change ourjobs, but we also have to think about how we need to understand what it's doing in order to do ourjobs, do you know i mean? we have to get much, much smarter about how we understand ai, what we know about al,
8:51 pm
how it's being used, who's using it. and we were talking about earlier about how newsrooms are using ai. if we flip it around and look at how ai has been using newsrooms, if you like, often without their permission. if what it is is essentially bots extracting data from all sorts of online sources, it's not clear if the tech companies have been paying news outlets to suck up all those years of news stories that have been paid for by whoever it might be, the bbc, the mail group, whoever, and then train their ais. or maybe it is becoming very clear that they haven't paid for this. they definitely haven't paid to do this, to build ai systems out of news publishers' data. but what does that mean? cos clearly there's a tension there. are these news organisations wising up to this now and saying, "you need to pay us"? how's that going to work? yeah, we reported a few weeks ago about basically all the biggest tech companies building ai models in talks with the biggest media
8:52 pm
publishers to kind of strike deals, proper financial deals about how they might be compensated for the use of news content because they have to scrape news websites in order to learn. and what happens when they, when you ask them a question about a news topic and they generate an answer? that's essentially generating journalism, but kind of sidestepping all of the sources that they used to train themselves. because i did read that the daily mail is looking at potentially suing google, taking legal action over the scraping of its news articles. yeah, you know, i think that... they're definitely wise to it. everybody�*s wise to it and as i said, there was, i think we named news corp, axel springer, the new york times, the guardian, all of these we know to have been in discussions around, you know, do theyjust pay you a cheque? do they build something for you? so there are definitely going to be partnerships that will have some kind of financial shape to them. i mean, we keep mentioning google, and obviously, there are lots of other organisations allegedly
8:53 pm
doing this as well. just to explain what you were saying, because i think for audiences, what this is going to mean is that the moment you might go into a search engine and you'll put in a question about something and it'll come up with a whole load of different articles that you can choose to read. and what we're saying is, in the future, it'll be a one—stop shop, the summary of all those articles, and that's what's different. jackson, is this something you're worried about? yeah, absolutely. this is the thing i'm most worried about as a digital publisher. i mean, from my point of view, we know that nine out of ten people use google, right? so, basically, all the internet search traffic goes through google right now still. and although google says that it has this idea that if it summarises something, it'll provide links so you can go deeper, i think we know that the behaviour of a searcher is not to try and go that deep. i don't think the second page of google is hardly ever clicked on. so we already know that summarising those articles is going to take away a lot of this traffic, and some of our digital publications are propped up by how much search traffic they get, right?
8:54 pm
like, i know that cnet's google traffic is like a big, big chunk of where we get our eyeballs from. it's predominantly through google. the business model gets broken by this summarising and what are the digital publications going to do? especially very, very new stuff. and that's why in a recent piece i argued like, we should not be allowing this to happen. we should have some sort of moratorium on how quickly these models can suck up data. i don't see any other way to prevent some of this from happening. i think we're coming towards the end and i would like to end by just looking to the future. you know, isjournalism doomed or is the debate about ai�*s application and news reporting actually a reminder of the value of human journalists? tom clarke from sky news. if we don't get this right, it could... i think there is a kind of existential crisis for information we're looking at. it's kind of what jackson was just touching on. microsoft, which invested,
8:55 pm
is investing $100 billion in openai, the company that generated chatgpt, gpt-4. they've already put that into being, so you can effectively use their search in the way that we were describing already. these tech companies are throwing everything at it and think on this, and this is what really troubles me, the more ai—generated content we put on the web without knowing whether it's accurate or not, the more data there is out there to scrape for the ms of the future. if we don't somehow manage to step in and separate what is true or human—generated, whether it's true or not, from the ai stuff, we get to a point where we're actually feeding the ai with ai—generated stuff and we might end up in a situation, in a very, very short order, because don't forget how much computing capacity goes into these, how much data they're able to scrape, we end up polluting the wellspring of the information that goes into the ms in the first place and we could be really, really stuck. so, i think there's a real crisis there, but there are also very important questions for journalists about how we can use these tools to make ourjobs better,
8:56 pm
assuming we survive this and continue gathering that information. i think to turn our back on al, say we have to just get rid of it, these tools could be so powerful for doing investigations, forfreeing up time, if it's done in the right way, for streamlining the work we do, getting stories out there faster. so i think ai tools have enormous potential in news that we mustn't overlook. thank you so much to you all for taking part in this media show, madhumita murgia from the ft, tom clarke from sky news, eliz mizon from the bristol cable, and jackson ryan from cnet. and, of course, thank you, everybody, who's been listening to the media show for now. thank you so much. goodbye. hello there. we've got a september heatwave building over the next few days with lots of dry weather, lots
8:57 pm
of sunshine and rising temperatures. we're looking at probably the hottest days wednesday, thursday, when we could see temperatures reach the low 30s. now, most of us had sunshine today. we've seen temperatures go past 26 degrees in a few areas. the one place that missed out on the sunshine and warm weather, the far north of scotland, where we had a bit of cloud and rain. and after a locally misty start to the day, we saw that sunshine really coming through across the board. this was penzance in cornwall, doing quite a good impression of the tropics, given the plants we've got across this part of cornwall. the plants we've got now, overnight tonight, we'll keep the clear weather, a few mist and fog patches are quite likely to develop. it's across shetland. it's across shetland, and we might see a few occasional patches of rain, but otherwise, it's a dry night — temperatures 12—16 degrees. the next few days weather—wise dominated by this big area of
8:58 pm
high pressure that's overalljust to the east of the uk, and it's this that will be bringing dry, sunny and very warm weather our way. now, from monday morning, you might start off seeing just a few early morning mist and fog patches. then they'll clear away within the first couple of hours of the morning, then sunshine across the board for most. the exception, again, the very far north of scotland might see a little bit of rain for shetland, but i think brighter weatherfor orkney, the hebrides and highland scotland as well. temperatures the highest across england and wales, reaching 27. but the warmest spots of scotland and northern ireland will also see temperatures into the mid—20s. so, for many, monday's going to be a warm day, and that warm theme continues for tuesday as well. as i say, for most of the uk, it's looking like another dry day with plenty of sunshine around, perhaps a little bit more cloud again across the northern isles of scotland, but otherwise, sunshine across the board. and for wednesday, probably a few more mist and fog patches around. some of those might linger around some of our coastal areas, particularly through the irish sea. but across the inland areas, we'll have lots of sunshine once the early morning mist and fog patches have cleared out of the way, and it continues to get hotter.
8:59 pm
temperatures could reach 30 degrees celsius, and that's not far away from the highest temperature we've recorded all year, which currently stands at 32.2. stays dry and sunny for most thursday, friday, saturday. it's only really until we get to sunday when we start to see a change to cooler, more unsettled weather conditions. live from london, this is bbc news. ukraine's military claims it's managed to break through key russian defences near the southeastern city of zaporizhzhia. calls for the government to "get a grip" on the crisis of crumbling concrete in school buildings. but the chancellor insists children will be kept safe. the labour leader, sir keir starmer, is expected to reshuffle his shadow cabinet tomorrow as mps return to westminster from their summer break.
9:00 pm
hello, i'm lukwesa burak. ukraine's generals are reporting significant progress in the counteroffensive against russia's invasion. they say their troops have breached the first line of russian defences in the south, near zaporizhzhia. general oleksandr tarnavskyi told the observer newspaper that extensive minefields had blocked troops for several weeks, leaving them vulnerable to russian shelling and drone attacks. but the general claimed that painstaking mine clearance has now allowed his forces to advance and they expect to face far weaker russian defences ahead. ukraine says its forces are making gains in the south of the country and has in recent weeks expanded its units towards the strategic town of tokmak, a logistical centre for russian forces. last week, ukraine's military said it had captured the village of robotyne in the zaporizhzhia region
17 Views
IN COLLECTIONS
BBC News Television Archive Television Archive News Search ServiceUploaded by TV Archive on