Thursday, November 28, 2024

The next evolution of AI begins with ours

Peer-Reviewed Publication

Cold Spring Harbor Laboratory

In a sense, each of us begins life ready for action. Many animals perform amazing feats soon after they’re born. Spiders spin webs. Whales swim. But where do these innate abilities come from? Obviously, the brain plays a key role as it contains the trillions of neural connections needed to control complex behaviors. However, the genome has space for only a small fraction of that information. This paradox has stumped scientists for decades. Now, Cold Spring Harbor Laboratory (CSHL) Professors Anthony Zador and Alexei Koulakov have devised a potential solution using artificial intelligence.

When Zador first encounters this problem, he puts a new spin on it. “What if the genome’s limited capacity is the very thing that makes us so smart?” he wonders. “What if it’s a feature, not a bug?” In other words, maybe we can act intelligently and learn quickly because the genome’s limits force us to adapt. This is a big, bold idea—tough to demonstrate. After all, we can’t stretch lab experiments across billions of years of evolution. That’s where the idea of the genomic bottleneck algorithm emerges.

In AI, generations don’t span decades. New models are born with the push of a button. Zador, Koulakov, and CSHL postdocs Divyansha Lachi and Sergey Shuvaev set out to develop a computer algorithm that folds heaps of data into a neat package—much like our genome might compress the information needed to form functional brain circuits. They then test this algorithm against AI networks that undergo multiple training rounds. Amazingly, they find the new, untrained algorithm performs tasks like image recognition almost as effectively as state-of-the-art AI. Their algorithm even holds its own in video games like Space Invaders. It’s as if it innately understands how to play.

Does this mean AI will soon replicate our natural abilities? “We haven’t reached that level,” says Koulakov. “The brain’s cortical architecture can fit about 280 terabytes of information—32 years of high-definition video. Our genomes accommodate about one hour. This implies a 400,000-fold compression technology cannot yet match.”

Nevertheless, the algorithm allows for compression levels thus far unseen in AI. That feature could have impressive uses in tech. Shuvaev, the study’s lead author, explains: “For example, if you wanted to run a large language model on a cell phone, one way [the algorithm] could be used is to unfold your model layer by layer on the hardware.”

Such applications could mean more evolved AI with faster runtimes. And to think, it only took 3.5 billion years of evolution to get here. 

 

AI Safety Institute launched as Korea’s AI Research Hub



Located at the Pangyo Global R&D Center, the AISI begins full operations under the leadership of a director and a three-section system



National Research Council of Science & Technology




The Ministry of Science and ICT (MSIT), headed by Minister Yoo Sang-im, held the launch ceremony for the "AI Safety Institute" (AISI) on Wednesday, November 27, at the Pangyo Global R&D Center.

At the "AI Seoul Summit"last May, leaders from 10 countries recognized safety as a key component of responsible AI innovation and emphasized the importance of establishing AI safety institutes and fostering global collaboration for safe AI. President Yoon Suk Yeol also expressed his commitment, stating, "We will work towards establishing an AI safety institute in Korea and actively participate in a global network to enhance AI safety." After thorough preparations regarding the institute's organization, budget, personnel, and functions, the AI Safety Institute has now been officially launched.

The AISI is a dedicated organization established within ETRI to systematically and professionally address various AI risks, including technological limitations, human misuse, and potential loss of control over AI. As Korea's hub for AI safety research, the AISI will facilitate collaborative research and information sharing among industry, academia, and research institutes in the field of AI safety. Furthermore, as a member of the "International Network of AI Safety Institutes" (comprising 10 countries, launched on November 21), the AISI is committed to taking a responsible role in strengthening global collaboration for safe AI. Through these efforts, the AISI aims to develop competitive technologies, nurture skilled professionals in the AI safety sector, and advance AI safety policies, including their development and refinement, based on scientific research data.

The launch ceremony brought togetherkey government officials, including Yoo Sang-im, Minister of Science and ICT; Yeom Jae-ho, Vice Chair of the National AI Committee; and Lee Kyung-woo, Presidential Secretary for AI and Digital. Over 40 prominent figures from the AI industry, academia, and research sectors also attended, such as Bae Kyung-hoon, Chief of LG AI Research; Oh Hye-yeon, Director of the KAIST AI Institute; Lee Eun-ju, Director of the Center for Trustworthy AI at Seoul National University; and Bang Seung-chan, President of the Electronics and Telecommunications Research Institute (ETRI).

At the event, Professor Yoshua Bengio, a globally renowned AI scholar and Global Advisor to the National AI Committee, congratulated the Korean government on establishing the AI Safety Institute in alignment with the Seoul Declaration. He emphasized the Institute's critical roles, including (1) researching and advancing risk assessment methodologies through industry collaboration, (2) supporting the development of AI safety requirements, and (3) fostering international cooperation to harmonize global AI safety standards. Additionally, the directors of AI safety institutes from the United States, the United Kingdom, and Japan delivered congratulatory speeches, stating, "We have high expectations for Korea’s AI Safety Institute" and emphasizing "the importance of global collaboration in AI safety."

Kim Myung-joo, the inaugural Director of the AISI, outlined the Institute's vision and operational plans during the ceremony. In his presentation, he stated, "The AISI will focus on evaluating potential risks that may arise from AI utilization, developingand disseminating policies and technologies to prevent and minimize these risks, and strengthening collaboration both domestically and internationally." Director Kim emphasized, "The AISI is not a regulatory body but a collaborative organization dedicated to supporting Korean AI companies by reducing risk factors that hinder their global competitiveness."

At the signing ceremony for the "Korea AI Safety Consortium" (hereinafter referred to as the "Consortium"), 24leading Korean organizations from industry, academia, and research sectors signed a Memorandum of Understanding (MOU) to promote mutual cooperation in AI safety policy research, evaluation, and R&D. The AISI and Consortium member organizations will jointly focus on key initiatives, including the research, development, and validation of an AI safety framework (risk identification, evaluation, and mitigation), policy research to align with international AI safety norms, and technological collaboration on AI safety. Moving forward, they plan to refine the Consortium's detailed research topics and operational strategies. The member organizations also presented their expertise in AI safety research and outlined their plans for Consortium activities, affirming their strong commitment to active collaboration with the AISI.

< Participating Organizations in the "AI Safety Consortium" >

Industry

 Naver (Future AI Center), KT (Responsible AI Center), Kakao (AI Safety), LG AI Research, SKT (AI Governance Task Force), Samsung Electronics, Konan Technology, Wrtn Technologies, ESTsoft, 42Maru, Crowdworks AI, Twelve Labs, Liner

Academia

▪Seoul National University (Center for Trustworthy AI), KAIST (AI Fairness Research Center), Korea University (School of Cybersecurity), Sungkyunkwan University (AI Reliability Research Center), Soongsil University (AI Safety Center), Yonsei University (AI Impact Research Center)

Research Institutes

▪Korea AISI, TTA (Center for Trustworthy AI), NIA (Department of AI Policy), KISDI (Department of Digital Society Strategy Research), IITP (AIí…±igital Convergence Division), SPRi (AI Policy Research Lab)

 

Minister Yoo Sang-im of the MSIT emphasized, "AI safety is a prerequisite for sustainable AI development and one of the greatest challenges that all of us in the AI field must tackle together." He noted, "In the short span of just one year since the AI Safety Summit in November 2023 and the AI Seoul Summit in May 2024, major countries such as the United States, the United Kingdom, Japan, Singapore, and Canada have established AI safety institutes, creating an unprecedentedly swift and systematic framework for international AI safety cooperation." Minister Yoo further emphasized, "By bringing together the research capabilities of industry, academia, and research institutes through the AISI, we will rapidly secure the technological and policy expertise needed to take a leading role in the global AI safety alliance. We will actively support the AISI's growth into a research hub representing the Asia-Pacific region in AI safety."

###

 

About Electronics and Telecommunications Research Institute (ETRI)

ETRI is a non-profit government-funded research institute. Since its foundation in 1976, ETRI, a global ICT research institute, has been making its immense effort to provide Korea a remarkable growth in the field of ICT industry. ETRI delivers Korea as one of the top ICT nations in the World, by unceasingly developing world’s first and best technologies.



New guidance for ensuring AI safety in clinical care published in JAMA by UTHealth Houston, Baylor College of Medicine researchers



University of Texas Health Science Center at Houston
Dean Sittig, PhD 

image: 

Dean Sittig, PhD

view more 

Credit: UTHealth Houston



As artificial intelligence (AI) becomes more prevalent in health care, organizations and clinicians must take steps to ensure its safe implementation and use in real-world clinical settings, according to an article co-written by Dean Sittig, PhD, professor with McWilliams School of Biomedical Informatics at UTHealth Houston and Hardeep Singh, MD, MPH, professor at Baylor College of Medicine.

The guidance was published today, Nov. 27, 2024, in the Journal of the American Medical Association.

“We often hear about the need for AI to be built safely, but not about how to use it safely in health care settings,” Sittig said. “It is a tool that has the potential to revolutionize medical care, but without safeguards in place, AI could generate false or misleading outputs that could potentially harm patients if left unchecked.” 

Drawing from expert opinion, literature reviews, and experiences with health IT use and safety assessment, Sittig and Singh developed a pragmatic approach for health care organizations and clinicians to monitor and manage AI systems.

“Health care delivery organizations will need to implement robust governance systems and testing processes locally to ensure safe AI and safe use of AI so that ultimately AI can be used to improve the safety of health care and patient outcomes,” Singh said. “All health care delivery organizations should check out these recommendations and start proactively preparing for AI now.”

Some of the recommended actions for health care organizations are listed below:

·       Review guidance published in high-quality, peer-reviewed journals and conduct rigorous real-world testing to confirm AI’s safety and effectiveness.

·       Establish dedicated committees with multidisciplinary experts to oversee AI system deployment and ensure adherence to safety protocols. Committee members should meet regularly to review requests for new AI applications, consider their safety and effectiveness before implementing them, and develop processes to monitor their performance.

·       Formally train clinicians on AI usage and risk, but also be transparent with patients when AI is part of their care decisions. This transparency is key to building trust and confidence in AI’s role in health care. 

·       Maintain a detailed inventory of AI systems and regularly evaluate them to identify and mitigate any risks.

·       Develop procedures to turn off AI systems should they malfunction, ensuring smooth transitions back to manual processes.

“Implementing AI into clinical settings should be a shared responsibility among health care providers, AI developers, and electronic health record vendors to protect patients,” Sittig said. “By working together, we can build trust and promote the safe adoption of AI in health care.”

Also providing input to the article were Robert Murphy, MD, associate professor and associate dean, and Debora Simmons, PhD, RN, assistant professor, both from the Department of Clinical and Health Informatics at McWilliams School of Biomedical Informatics; and Trisha Flanagan, RN, MSN.

 

No comments:

Post a Comment