CBS News
Flooding causes Rapidan Dam near Mankato to fail; nearby residents urged to evacuate
RAPIDIAN, Minn. — Flooding on the Blue Earth River caused the Rapidan Dam to fail on Monday morning. Le Sueur County residents in low-lying areas of the Minnesota River Valley are advised to evacuate, according to the county’s emergency management office.
The National Weather Service has issued a flash flood warning for areas downstream until 4:30 p.m.
The Blue Earth County Sheriff’s Office says debris started accumulating on Sunday at the 114-year-old dam, located about 10 miles southwest of Mankato, placing it “in imminent failure condition.”
Dam operators alerted officials at about 10:36 a.m. Monday the river had significantly eroded earth around the west side of the dam on the south side of Mankato, the National Weather Service says.
Pieces of an Xcel Energy building and other county buildings are adding to the debris in the water.
Officials also say the County Road 33 Bridge and County Road 90 Bridge are being closely monitored for debris passing downstream and may be closed if needed.
The Blue Earth County Sheriff’s Office says they are also aware of power outages and are working with Xcel to restore them.
The failure is expected to cause the Minnesota River at Mankato to crest just below major flood state on Tuesday morning, the National Weather Service says.
The Blue Earth River begins in northern Iowa and meets with the river’s west branch in Faribault County in southern Minnesota, according to the Minnesota Pollution Control Agency. From there, it flows 108 miles north past the cities of Blue Earth, Winnebago, and Vernon Center to Mankato, where it enters the Minnesota River.
There are 21 cities in that watershed, of which Mankato and Fairmont are the largest.
This is a developing story. Stay with WCCO.com for more.
CBS News
Teen victim of AI-generated “deepfake pornography” urges Congress to pass “Take It Down Act”
Anna McAdams has always kept a close eye on her 15-year-old daughter Elliston Berry’s life online. So it was hard to come to terms with what happened 15 months ago on the Monday morning after Homecoming in Aledo, Texas.
A classmate took a picture from Elliston’s Instagram, ran it through an artificial intelligence program that appeared to remove her dress and then sent around the digitally altered image on Snapchat.
“She came into our bedroom crying, just going, ‘Mom, you won’t believe what just happened,'” McAdams said.
Last year, there were more than 21,000 deepfake pornographic videos online — up more than 460% over the year prior. The manipulated content is proliferating on the internet as websites make disturbing pitches — like one service that asks, “Have someone to undress?”
“I had PSAT testing and I had volleyball games,” Elliston said. “And the last thing I need to focus and worry about is fake nudes of mine going around the school. Those images were up and floating around Snapchat for nine months.”
In San Francisco, Chief Deputy City Attorney Yvonne Mere was starting to hear stories similar to Elliston’s — which hit home.
“It could have easily been my daughter,” Mere said.
The San Francisco City Attorney’s office is now suing the owners of 16 websites that create “deepfake nudes,” where artificial intelligence is used to turn non-explicit photos of adults and children into pornography.
“This case is not about tech. It’s not about AI. It’s sexual abuse,” Mere said.
These 16 sites had 200 million visits in just the first six months of the year, according to the lawsuit.
City Attorney David Chiu says the 16 sites in the lawsuit are just the start.
“We’re aware of at least 90 of these websites. So this is a large universe and it needs to be stopped,” Chiu said.
Republican Texas Sen. Ted Cruz is co-sponsoring another angle of attack with Democratic Minnesota Sen. Amy Klochubar. The Take It Down Act would require social media companies and websites to remove non-consensual, pornographic images created with AI.
“It puts a legal obligation on any tech platform — you must take it down and take it down immediately,” Cruz said.
The bill passed the Senate this month and is now attached to a larger government funding bill awaiting a House vote.
In a statement, a spokesperson for Snap told CBS News: “We care deeply about the safety and well-being of our community. Sharing nude images, including of minors, whether real or AI-generated, is a clear violation of our Community Guidelines. We have efficient mechanisms for reporting this kind of content, which is why we’re so disheartened to hear stories from families who felt that their concerns went unattended. We have a zero tolerance policy for such content and, as indicated in our latest transparency report, we act quickly to address it once reported.”
Elliston says she’s now focused on the present and is urging Congress to pass the bill.
“I can’t go back and redo what he did, but instead, I can prevent this from happening to other people,” Elliston said.
CBS News
Lawmakers target AI-generated “deepfake pornography”
Be the first to know
Get browser notifications for breaking news, live events, and exclusive reporting.
CBS News
A look inside the U.S.-Mexico border
Be the first to know
Get browser notifications for breaking news, live events, and exclusive reporting.