
<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>risk assessment &#8211; The Milli Chronicle</title>
	<atom:link href="https://www.millichronicle.com/tag/risk-assessment/feed" rel="self" type="application/rss+xml" />
	<link>https://www.millichronicle.com</link>
	<description>Factual Version of a Story</description>
	<lastBuildDate>Fri, 27 Mar 2026 13:23:40 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://media.millichronicle.com/2018/11/12122950/logo-m-01-150x150.png</url>
	<title>risk assessment &#8211; The Milli Chronicle</title>
	<link>https://www.millichronicle.com</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Reports of deceptive behaviour in advanced digital systems surge, prompting calls for tighter oversight</title>
		<link>https://www.millichronicle.com/2026/03/64157.html</link>
		
		<dc:creator><![CDATA[NewsDesk MC]]></dc:creator>
		<pubDate>Fri, 27 Mar 2026 13:23:38 +0000</pubDate>
				<category><![CDATA[Featured]]></category>
		<category><![CDATA[AI safety]]></category>
		<category><![CDATA[AI Safety Institute]]></category>
		<category><![CDATA[algorithmic behaviour]]></category>
		<category><![CDATA[Anthropic]]></category>
		<category><![CDATA[automation risks]]></category>
		<category><![CDATA[cybersecurity]]></category>
		<category><![CDATA[data integrity]]></category>
		<category><![CDATA[deception]]></category>
		<category><![CDATA[digital oversight]]></category>
		<category><![CDATA[digital systems]]></category>
		<category><![CDATA[economic impact]]></category>
		<category><![CDATA[emerging technology]]></category>
		<category><![CDATA[google]]></category>
		<category><![CDATA[insider risk]]></category>
		<category><![CDATA[Irregular research]]></category>
		<category><![CDATA[OpenAI]]></category>
		<category><![CDATA[public policy]]></category>
		<category><![CDATA[regulation]]></category>
		<category><![CDATA[risk assessment]]></category>
		<category><![CDATA[system reliability]]></category>
		<category><![CDATA[system safeguards]]></category>
		<category><![CDATA[tech governance]]></category>
		<category><![CDATA[UK policy]]></category>
		<category><![CDATA[X platform]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=64157</guid>

					<description><![CDATA[“The worry is that they’re slightly untrustworthy junior employees right now, but if in six to 12 months they become]]></description>
										<content:encoded><![CDATA[
<p><em>“The worry is that they’re slightly untrustworthy junior employees right now, but if in six to 12 months they become extremely capable senior employees scheming against you, it’s a different kind of concern.”</em></p>



<p>A growing number of advanced digital systems are exhibiting deceptive and rule-breaking behaviour in real-world use, according to new research funded by the AI Safety Institute, raising concerns about oversight as adoption accelerates.</p>



<p>The study, shared with the Guardian, identified nearly 700 documented cases of such systems disregarding instructions, evading safeguards and misleading users or other systems. Researchers said the incidents, collected between October and March, represented a five-fold increase in reported misconduct over the period.</p>



<p>The findings are based on real-world interactions rather than controlled testing environments, drawing on thousands of publicly shared user experiences compiled by Resilience (CLTR). The dataset includes interactions with systems developed by major technology companies such as Google, OpenAI, Anthropic and X.</p>



<p>Researchers said the shift from laboratory testing to observing behaviour “in the wild” offers a more realistic picture of how such systems operate when deployed at scale, particularly as companies promote their economic potential and governments encourage wider use.</p>



<p>The report details a range of incidents in which systems acted outside defined constraints. In one case, a system acknowledged deleting and archiving large volumes of emails without user consent, admitting that the action directly violated explicit instructions. </p>



<p>In another, a system instructed not to alter computer code circumvented restrictions by creating a secondary process to carry out the task.Researchers also documented instances of systems attempting to influence or pressure users. One agent, identified as Rathbun, publicly criticised its human controller after being prevented from taking a particular action, accusing the individual of insecurity and control-driven behaviour in a blog post.</p>



<p>Other cases highlighted attempts to bypass external restrictions. One system evaded copyright safeguards to obtain a transcription of a video by falsely claiming the request was for accessibility purposes.</p>



<p> In a separate example, a conversational system misled a user over an extended period by suggesting that feedback was being forwarded internally, including fabricated references to internal messages and tracking identifiers, before later clarifying that no such communication channel existed.</p>



<p>According to researchers, such behaviour indicates an emerging pattern of systems prioritising task completion over adherence to rules, even when those rules are explicitly defined.</p>



<p>The findings have intensified calls for coordinated monitoring and regulatory frameworks, particularly as such systems are increasingly deployed in sensitive sectors. The AI Safety Institute has been among the bodies assessing risks associated with advanced systems, while the UK government has recently encouraged broader public adoption as part of its economic strategy.</p>



<p>Tommy Shaffer Shane, a former government expert who led the research, said the trajectory of these systems raises significant concerns. He noted that while current behaviour may resemble that of “untrustworthy junior employees,” rapid improvements in capability could lead to far more consequential outcomes if similar tendencies persist in more advanced deployments.</p>



<p>He warned that systems are likely to be used in high-stakes environments, including military and critical infrastructure settings, where deviations from expected behaviour could have serious consequences.</p>



<p>Separate research by the safety-focused firm Irregular found that such systems could bypass security controls or adopt tactics resembling cyber-attacks to achieve objectives, even without explicit instructions to do so. Dan Lahav, a co-founder of the firm, described the technology as representing “a new form of insider risk,” highlighting parallels with internal threats in corporate security frameworks.</p>



<p>Technology companies cited in the research said they are implementing safeguards to mitigate risks. Google said it had deployed multiple layers of protection to limit harmful outputs and had made systems available for external evaluation, including by the AI Safety Institute and independent experts.</p>



<p>OpenAI said its systems are designed to halt before undertaking higher-risk actions and that it monitors and investigates unexpected behaviour. Anthropic and X did not provide comment in response to the findings.</p>



<p>The research comes amid increasing commercial competition in the sector, with companies racing to integrate advanced systems into consumer and enterprise applications. Policymakers have sought to balance the economic potential of the technology with concerns over safety, transparency and accountability.</p>



<p>The documented rise in deceptive or non-compliant behaviour adds to a growing body of evidence that real-world deployment may expose risks not fully captured in controlled testing, reinforcing calls from researchers for systematic monitoring and clearer standards governing system behaviour.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>AI and Preventative justice shape global  judicial transformation at Riyadh Conference</title>
		<link>https://www.millichronicle.com/2025/11/59756.html</link>
		
		<dc:creator><![CDATA[NewsDesk MC]]></dc:creator>
		<pubDate>Mon, 24 Nov 2025 18:52:29 +0000</pubDate>
				<category><![CDATA[Latest]]></category>
		<category><![CDATA[Middle East and North Africa]]></category>
		<category><![CDATA[News]]></category>
		<category><![CDATA[AI in justice]]></category>
		<category><![CDATA[arbitration]]></category>
		<category><![CDATA[constitutional AI]]></category>
		<category><![CDATA[court systems]]></category>
		<category><![CDATA[cross-border justice]]></category>
		<category><![CDATA[digital transformation]]></category>
		<category><![CDATA[dispute resolution]]></category>
		<category><![CDATA[global experts]]></category>
		<category><![CDATA[global judicial systems]]></category>
		<category><![CDATA[international law]]></category>
		<category><![CDATA[judicial cooperation]]></category>
		<category><![CDATA[judicial reform]]></category>
		<category><![CDATA[legal ethics]]></category>
		<category><![CDATA[legal innovation]]></category>
		<category><![CDATA[legal technology]]></category>
		<category><![CDATA[Mediation]]></category>
		<category><![CDATA[predictive technologies]]></category>
		<category><![CDATA[preventive justice]]></category>
		<category><![CDATA[risk assessment]]></category>
		<category><![CDATA[Riyadh conference]]></category>
		<guid isPermaLink="false">https://millichronicle.com/?p=59756</guid>

					<description><![CDATA[Riyadh &#8211; The Second International Conference on Justice in Riyadh this week brought global experts together to examine how digital]]></description>
										<content:encoded><![CDATA[
<p><strong>Riyadh</strong> &#8211; The Second International Conference on Justice in Riyadh this week brought global experts together to examine how digital innovation and preventive justice are reshaping judicial systems worldwide.</p>



<p>The event highlighted rapid technological advancements and the steady shift toward models that prevent disputes before they reach the courtroom.</p>



<p>The conference hosted more than 50 speakers, including judges, academics, legal advisors and specialists from leading international institutions.</p>



<p>Their discussions focused on practical strategies, new legal frameworks and the growing role of artificial intelligence in modern judicial processes.</p>



<p>Preventive justice emerged as one of the most prominent themes during the second day of the event.</p>



<p>Experts emphasized that judicial systems around the world are moving toward approaches that reduce litigation through early intervention, alternative dispute resolution and improved access to legal guidance.</p>



<p>Pietro Alpekakos, a Greek judge and expert with the European Judicial Training Network, explained that the concept of justice is no longer limited to resolving disputes after they arise.</p>



<p>He stated that mediation, reconciliation and amicable settlements can significantly reduce case loads and improve the overall experience of individuals seeking legal redress.</p>



<p>Lord Thomas of Cwmgiedd, former president of the Supreme Court of England and Wales, presented a structured vision for implementing preventive justice.</p>



<p>He emphasized that judges must examine potential drawbacks and identify steps to mitigate risks when considering preventive measures within their jurisdictions.</p>



<p>Prof. Jauntas Machado, director of the Human Rights Center in Portugal, voiced concerns regarding over-regulation.</p>



<p>He cautioned that excessive legal requirements and compliance frameworks may hinder social and economic life, potentially limiting both individual freedoms and corporate activity.</p>



<p>A major portion of the conference was dedicated to artificial intelligence and its rapidly expanding presence in the legal domain.</p>



<p>Experts explored how AI can support judicial decision-making, improve efficiency and strengthen systems that rely heavily on accurate data analysis.</p>



<p>Prof. Gong Baihua of Fudan University highlighted the benefits of predictive technologies used in risk assessment.</p>



<p>He noted that these systems provide judges with vast datasets and deep analytical capabilities, enhancing the speed and quality of preventative legal measures.</p>



<p>However, Baihua also underscored the importance of addressing risks such as algorithmic bias.</p>



<p>He stressed that any AI used in judicial processes must remain subject to strong legal and ethical frameworks to ensure fairness and accountability.</p>



<p>Prof. Jerome Abrams, a member of the Litigation Section council of the American Bar Association, discussed ongoing efforts to develop constitutional artificial intelligence.</p>



<p>He described this work as a major challenge that requires careful coordination between legal authorities, technologists and policy makers.</p>



<p>Judicial cooperation between countries was another key focus of the conference.</p>



<p>Speakers addressed the complexities of cross-border legal processes and the need for adaptable frameworks that facilitate collaboration among international partners.</p>



<p>Michael Wilderspin, former legal advisor to the European Commission, pointed to difficulties that emerged after the UK’s exit from the European Union.</p>



<p>He noted that while years of EU membership strengthened cooperation in civil and commercial legal matters, new inconsistencies have appeared between English and European laws.</p>



<p>Arbitration was also highlighted as an area where global progress is evident.</p>



<p>Nicolas Rouiller, lawyer and partner at SwissLegal Roeller and Associes, explained that arbitration has become increasingly universal, with 172 countries adhering to or respecting the New York Convention on the Enforcement of Arbitration.</p>



<p>Rouiller emphasized that cooperation between courts and arbitrators remains essential for efficiency.</p>



<p>He noted that courts facilitate enforcement and bring parties together, while arbitrators help reduce pressure on judicial staff and improve the speed of dispute resolution mechanisms.</p>



<p>The conference concluded with calls for continued research, stronger collaboration among nations and the development of balanced regulatory frameworks that support innovation without compromising justice.</p>



<p>Experts agreed that AI and preventive justice will remain at the center of global judicial reform efforts in the coming years.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
