Fokke & Sukke

F & S

Behance Featured Projects

The latest projects featured on the Behance

Veenland Pitch


Can't Recall the Name, But the Face Is Familiar

gecko47 has added a photo to the pool:

Can't Recall the Name, But the Face Is Familiar

I've scoured my Dragonflies Of Australia field guide but nothing seems to fit. I'm sure this is a simple task for some Damselfly expert out there!
Photographed in Brisbane in Minnippi Parkland.

A Meeting

gecko47 has added a photo to the pool:

A Meeting

A number of Lewin's Honeyeaters were finding a corner of Minnippi Parkland productive in their search for food.

Bush

Markus Branse has added a photo to the pool:

Bush

Bush, Litchfield Nationalpark, Northern Territory, Australia

Bush

Markus Branse has added a photo to the pool:

Bush

Bush, Litchfield Nationalpark, Northern Territory, Australia

Outback gem

irwinm16 has added a photo to the pool:

Outback gem

Slashdot

News for nerds, stuff that matters

Actors Who Sold AI Avatars Stuck In Black Mirror-Esque Dystopia

Some actors who sold their likenesses to AI video companies like Synthesia now regret the decision, after finding their digital avatars used in misleading, embarrassing, or politically charged content. Ars Technica reports: Among them is a 29-year-old New York-based actor, Adam Coy, who licensed rights to his face and voice to a company called MCM for one year for $1,000 without thinking, "am I crossing a line by doing this?" His partner's mother later found videos where he appeared as a doomsayer predicting disasters, he told the AFP. South Korean actor Simon Lee's AI likeness was similarly used to spook naive Internet users but in a potentially more harmful way. He told the AFP that he was "stunned" to find his AI avatar promoting "questionable health cures on TikTok and Instagram," feeling ashamed to have his face linked to obvious scams. [...]

Even a company publicly committed to ethically developing AI avatars and preventing their use in harmful content like Synthesia can't guarantee that its content moderation will catch everything. A British actor, Connor Yeates, told the AFP that his video was "used to promote Ibrahim Traore, the president of Burkina Faso who took power in a coup in 2022" in violation of Synthesia's terms. [...] Yeates was paid about $5,000 for a three-year contract with Synthesia that he signed simply because he doesn't "have rich parents and needed the money." But he likely couldn't have foreseen his face being used for propaganda, as even Synthesia didn't anticipate that outcome.

Others may not like their AI avatar videos but consider the financial reward high enough to make up for the sting. Coy confirmed that money motivated his decision, and while he found it "surreal" to be depicted as a con artist selling a dystopian future, that didn't stop him from concluding that "it's decent money for little work." Potentially improving the climate for actors, Synthesia is forming a talent program that it claims will give actors a voice in decision-making about AI avatars. "By involving actors in decision-making processes, we aim to create a culture of mutual respect and continuous improvement," Synthesia's blog said.

Read more of this story at Slashdot.

Study Finds 50% of Workers Use Unapproved AI Tools

An anonymous reader quotes a report from SecurityWeek: An October 2024 study by Software AG suggests that half of all employees are Shadow AI users, and most of them wouldn't stop even if it was banned. The problem is the ease of access to AI tools, and a work environment that increasingly advocates the use of AI to improve corporate efficiency. It is little wonder that employees seek their own AI tools to improve their personal efficiency and maximize the potential for promotion. It is frictionless, says Michael Marriott, VP of marketing at Harmonic Security. 'Using AI at work feels like second nature for many knowledge workers now. Whether it's summarizing meeting notes, drafting customer emails, exploring code, or creating content, employees are moving fast.' If the official tools aren't easy to access or if they feel too locked down, they'll use whatever's available which is often via an open tab on their browser.

There is almost also never any malicious intent (absent, perhaps, the mistaken employment of rogue North Korean IT workers); merely a desire to do and be better. If this involves using unsanctioned AI tools, employees will likely not disclose their actions. The reasons may be complex but combine elements of a reluctance to admit that their efficiency is AI assisted rather than natural, and knowledge that use of personal shadow AI might be discouraged. The result is that enterprises often have little knowledge of the extent of Shadow IT, nor the risks it may present. According to an analysis from Harmonic, ChatGPT is the dominant gen-AI model used by employees, with 45% of data prompts originating from personal accounts (such as Gmail). Image files accounted for 68.3%. The report also notes that 7% of empmloyees were using Chinese AI models like DeepSeek, Baidu Chat and Qwen.

"Overall, there has been a slight reduction in sensitive prompt frequency from Q4 2024 (down from 8.5% to 6.7% in Q1 2025)," reports SecurityWeek. "However, there has been a shift in the risk categories that are potentially exposed. Customer data (down from 45.8% to 27.8%), employee data (from 26.8% to 14.3%) and security (6.9% to 2.1%) have all reduced. Conversely, legal and financial data (up from 14.9% to 30.8%) and sensitive code (5.6% to 10.1%) have both increased. PII is a new category introduced in Q1 2025 and was tracked at 14.9%."

Read more of this story at Slashdot.

MetaFilter

The past 24 hours of MetaFilter

Where Discoveries Began

The National Science Foundation has updated it's priorities The U.S. National Science Foundation (NSF) was established in 1950 to promote the progress of science, advance the national health, prosperity and welfare, and secure the national defense. In response to the new administration they have updated their priorities. It's not good.

Included in the updates are: Awards that are not aligned with NSF's priorities have been terminated, including but not limited to those on diversity, equity, and inclusion (DEI). and Investigators wishing to expand participation in STEM for women and underrepresented groups, must ensure that all outreach, recruitment, or participatory activities in NSF projects are open and available to all Americans. So, not underrepresented groups. and NSF will not support research with the goal of combating "misinformation," "disinformation," and "malinformation". If this makes you mad, you know who to call.