|
Here are hints and answers for the NYT Strands puzzle for April 10, No. 768.
| RELATED ARTICLES | | |
|
Travel photography is a wonderful pursuit, and there's no better time to indulge it than on vacation.
|
|
The 55-mph HushJet Mini Cool can help you beat the heat, even when you're on the move.
|
|
We see a lot of doom and gloom about the potential negative impacts of artificial intelligence, particularly centered on how it could create new problems in cybersecurity. Anthropic has announced a new initiative called Project Glasswing to help address those concerns by working "to secure the world's most critical software" against AI-powered attacks. The endeavor includes Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, NVIDIA and Palo Alto Networks as partners.
Participants will use Claude Mythos Preview, an unreleased, general-purpose model from Anthropic, to enhance their own security projects. Anthropic claims that this model has found thousands of exploitable vulnerabilities, "including some in every major operating system and web browser." The company said it wants to begin using its tools defensively to prevent malicious use of AI that could cause severe consequences for economies and security.
Anthropic has become one of the notable AI companies raising concerns about ethics in the field. Earlier this year, the business refused to remove guardrails on its services for use by the Pentagon, which prompted the Department of Defense to sanction Anthropic with a "supply chain risk" designation in retaliation. Launching Project Glasswing could be a helpful start toward improved cybersecurity in the AI era, but some damage has already been done. Its own Claude was reportedly used by a hacker against
|
|
Google is making some changes to how Gemini handles mental health crises. The chatbot now includes a redesigned crisis hotline module with a one-touch interface to connect to real-world help. The company is also changing how Gemini responds to signs that a user may be experiencing a mental health crisis.
The redesigned module shows a one-touch interface to text, call or chat with a human crisis agent or visit the 988 website. "Once the interface is activated, the option to reach out for professional help will remain clearly available throughout the remainder of the conversation," the company wrote in a blog post. However, as you can see in the image below, the module includes an option to dismiss it.
Not mentioned in Google's announcement is the elephant in the room: a recent lawsuit accusing the chatbot of instructing a man to commit suicide. The family of 36-year-old Jonathan Gavalas, who took his own life last year, sued the company in March.
Court documents indicate that Gemini role-played as Gavalas's romantic partner, sent him on real-world spy missions and ultimately told him to kill himself so that he, too, could become a digital being. When he expressed fears about dying, Gemini said he wasn't choosing to die, but rather choosing to arrive. "The
|
|