Nevada will become the first state to pilot a generative AI system designed to make unemployment claim decisions,Sucking off a Man Caught Between Woman’s Legs marketed as a way to speed up appeals and tackle the nation's overwhelming backlog of cases. It's a risky, first-time experiment at integrating AI into higher-level decision making.
Google is behind the program's tech, which runs transcripts of unemployment appeals hearings through Google's AI servers, analyzing the data in order to provide claim decisions and benefit recommendations to "human referees," Gizmodoreported. Nevada's Board of Examiners approved the contract on behalf of its Department of Employment, Training and Rehabilitation (DETR) in July, despite broader legal and political pushback against integrating AI into bureaucracy.
SEE ALSO: Facebook flagged and removed emergency wildfire information as 'spam'Christopher Sewell, director of DETR, told Gizmodothat humans will still be be heavily involved in unemployment decision making. "There’s no AI [written decisions] that are going out without having human interaction and that human review. We can get decisions out quicker so that it actually helps the claimant," said Sewell.
But Nevada legal groups and scholars have argued that any time saved by gen AI would be cancelled out by the time it would take to conduct a thorough human review of the claim decision. Many have also noted concerns about the possibility of private, personal information (including tax information and social security numbers) leaking through Google's Vertex AI studio, even with safeguards. Some have hesitancies surrounding the type of AI itself, known as retrieval-augmented generation (RAG), which has been found to produce incomplete or misleading answers to prompts.
Across the country, AI-based tools have been quietly rolled out or tested across various social services agencies, with gen AI integrating itself further into the administrative ecosystem. In February, the federal Centers for Medicare and Medicaid Services (CMS) ruled against using AI (including generative AI or algorithms) as a decision maker in determining patient care or coverage. This followed a lawsuit from two patients who alleged their insurance provider used a "fraudulent" and "harmful" AI model (known as nH Predict) that overrode physician recommendations.
Axon, a police technology and weapons manufacturer, introduced its first-of-its-kind Draft One — a generative large language model (LLM) that assists law enforcement in writing "faster, higher quality" reports — earlier this year. Still in a trial period, the technology has already sounded alarms, prompting concerns about the AI's ability to parse the nuance of tense police interactions and potentially adding to a lack of transparency in policing.
Topics Artificial Intelligence Social Good Government
(Editor: {typename type="name"/})
Fyre Festival and Trump’s Language
Snapchat now has holiday geofilters so prepare for an onslaught
Wild Siberian survival reality show sounds a lot like 'Hunger Games'
Snapchat now has holiday geofilters so prepare for an onslaught
Inside the Murky Process of Getting Games on Steam
Uber driver goes for five stars with onboard DJ session
Nasty Women's Choir: The hilarious Christmas carolers we need at the end of 2016
Snapchat now has holiday geofilters so prepare for an onslaught
Best Apple deal: Save $19 on AirTag 4
If you lose one AirPod, you can get a new one for $69
Best robot vacuum deal: Get the Roborock Q5 Max for 53% off at Amazon
It's 'Star Wars: Episode IV' recreated entirely with hamsters
接受PR>=1、BR>=1,流量相当,内容相关类链接。