Nikhil, Author at MarkTechPost https://www.marktechpost.com/author/nikhil0980/ An Artificial Intelligence News Platform Fri, 27 Dec 2024 20:02:40 +0000 en-US hourly 1 https://wordpress.org/?v=6.7.1 https://www.marktechpost.com/wp-content/uploads/2022/04/cropped-Favicon-512-x-512-1-1-32x32.png Nikhil, Author at MarkTechPost https://www.marktechpost.com/author/nikhil0980/ 32 32 127842392 Google DeepMind Introduces Differentiable Cache Augmentation: A Coprocessor-Enhanced Approach to Boost LLM Reasoning and Efficiency https://www.marktechpost.com/2024/12/27/google-deepmind-introduces-differentiable-cache-augmentation-a-coprocessor-enhanced-approach-to-boost-llm-reasoning-and-efficiency/ https://www.marktechpost.com/2024/12/27/google-deepmind-introduces-differentiable-cache-augmentation-a-coprocessor-enhanced-approach-to-boost-llm-reasoning-and-efficiency/#respond Fri, 27 Dec 2024 20:02:30 +0000 https://www.marktechpost.com/?p=66757 Large language models (LLMs) are integral to solving complex problems across language processing, mathematics, and reasoning domains. Enhancements in computational techniques focus on enabling LLMs to process data more effectively, generating more accurate and contextually relevant responses. As these models become complex, researchers strive to develop methods to operate within fixed computational budgets without sacrificing […]

The post Google DeepMind Introduces Differentiable Cache Augmentation: A Coprocessor-Enhanced Approach to Boost LLM Reasoning and Efficiency appeared first on MarkTechPost.

]]>

Large language models (LLMs) are integral to solving complex problems across language processing, mathematics, and reasoning domains. Enhancements in computational techniques focus on enabling LLMs to process data more effectively, generating more accurate and contextually relevant responses. As these models become complex, researchers strive to develop methods to operate within fixed computational budgets without sacrificing performance.

One major challenge in optimizing LLMs is their inability to effectively reason across multiple tasks or perform computations beyond their pre-trained architecture. Current methods for improving model performance involve generating intermediate steps during task processing, often at the cost of increased latency and computational inefficiency. This limitation hampers their ability to perform complex reasoning tasks, particularly those requiring longer dependencies or higher accuracy in predictions.

Researchers have explored methods like Chain-of-Thought (CoT) prompting, which guides LLMs to reason step by step. While effective in some cases, CoT relies on sequential processing of intermediate reasoning steps, leading to slower computation times. KV-cache compression has also been proposed to reduce memory usage but does little to improve reasoning capabilities. These approaches, though valuable, underscore the need for a method that combines efficiency with enhanced reasoning ability.

Researchers from Google DeepMind have introduced a method called Differentiable Cache Augmentation. This technique uses a trained coprocessor to augment the LLM’s key-value (kv) cache with latent embeddings, enriching the model’s internal memory. The key innovation lies in keeping the base LLM frozen while training the coprocessor, which operates asynchronously. The researchers designed this method to enhance reasoning capabilities without increasing the computational burden during task execution.

The methodology revolves around a three-stage process. First, the frozen LLM generates a kv-cache from an input sequence, encapsulating its internal representation. This kv-cache is passed to the coprocessor, which processes it with additional trainable soft tokens. Not tied to specific words, these tokens act as abstract prompts for generating latent embeddings. Once processed, the augmented kv-cache is fed back into the LLM, enabling it to generate contextually enriched outputs. This asynchronous operation ensures the coprocessor’s enhancements are applied efficiently without delaying the LLM’s primary functions. Training the coprocessor is conducted using a language modeling loss, focusing solely on its parameters while preserving the integrity of the frozen LLM. This targeted approach allows for scalable and effective optimization.

Performance evaluations demonstrated significant improvements. The method was tested on the Gemma-2 2B model, achieving considerable results across various benchmarks. For instance, on the reasoning-intensive GSM8K dataset, accuracy improved by 10.05% when 64 latent embeddings were used. Similarly, MMLU performance increased by 4.70% under the same configuration. These enhancements underscore the model’s ability to perform better on complex reasoning tasks. Further, perplexity reductions were observed at multiple token positions. For example, perplexity decreased by 3.94% at position one and 1.20% at position 32 when 64 latent embeddings were applied, showcasing the model’s improved prediction capabilities over longer sequences.

Further analysis showed that the augmentation’s effectiveness scales with the number of latent embeddings. For GSM8K, accuracy rose incrementally with additional embeddings, from 1.29% with four embeddings to the peak improvement of 10.05% with 64 embeddings. Similar trends were observed in other benchmarks like ARC and MATH, indicating the broader applicability of this method. The researchers confirmed that their approach consistently outperformed baseline models without task-specific fine-tuning, demonstrating its robustness and adaptability.

This work represents a significant step forward in enhancing LLMs’ reasoning capabilities. By introducing an external coprocessor to augment the kv-cache, the researchers from Google DeepMind have created a method that improves performance while maintaining computational efficiency. The results highlight the potential for LLMs to tackle more complex tasks, paving the way for further exploration into modular enhancements and scalable reasoning systems. This breakthrough underscores the importance of continual innovation in AI to meet the growing demands of reasoning-intensive applications.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Google DeepMind Introduces Differentiable Cache Augmentation: A Coprocessor-Enhanced Approach to Boost LLM Reasoning and Efficiency appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/27/google-deepmind-introduces-differentiable-cache-augmentation-a-coprocessor-enhanced-approach-to-boost-llm-reasoning-and-efficiency/feed/ 0 66757
This Research from Amazon Explores Step-Skipping Frameworks: Advancing Efficiency and Human-Like Reasoning in Language Models https://www.marktechpost.com/2024/12/26/this-research-from-amazon-explores-step-skipping-frameworks-advancing-efficiency-and-human-like-reasoning-in-language-models/ https://www.marktechpost.com/2024/12/26/this-research-from-amazon-explores-step-skipping-frameworks-advancing-efficiency-and-human-like-reasoning-in-language-models/#respond Fri, 27 Dec 2024 00:15:09 +0000 https://www.marktechpost.com/?p=66734 The pursuit of enhancing artificial intelligence (AI) capabilities is significantly influenced by human intelligence, particularly in reasoning and problem-solving. Researchers aim to create language models that emulate human-like behaviors, such as optimizing reasoning processes. This involves exploring how models can transition from detailed, step-by-step solutions to more efficient methods by selectively skipping steps, a hallmark […]

The post This Research from Amazon Explores Step-Skipping Frameworks: Advancing Efficiency and Human-Like Reasoning in Language Models appeared first on MarkTechPost.

]]>

The pursuit of enhancing artificial intelligence (AI) capabilities is significantly influenced by human intelligence, particularly in reasoning and problem-solving. Researchers aim to create language models that emulate human-like behaviors, such as optimizing reasoning processes. This involves exploring how models can transition from detailed, step-by-step solutions to more efficient methods by selectively skipping steps, a hallmark of human expertise. These advancements contribute to achieving artificial general intelligence (AGI) with improved efficiency and task-solving capabilities.

A key challenge in AI is the models’ inability to replicate humans’ selective approach to skipping redundant steps during problem-solving. Humans develop this skill through practice, which allows them to reduce cognitive effort and focus on more complex aspects of a problem. Current language models lack this ability, adhering strictly to detailed processes even when simpler, equally effective solutions exist. Developing models incorporating such step-skipping behavior can enhance their efficiency and generalization abilities across various tasks.

Traditional training methods for language models involve step-by-step reasoning, relying on detailed datasets. Techniques such as chain-of-thought prompting encourage sequential solutions but do not address step skipping. As a result, while these models excel in solving problems comprehensively, they fail to demonstrate the efficiency observed in human experts. This limitation presents an opportunity to refine model training approaches to integrate more flexible reasoning capabilities.

Researchers from institutions like Fudan University, UC Santa Barbara, Shanghai AI Laboratory, Westlake University, and Amazon AWS AI developed a novel framework to address this. This approach introduces controlled training environments where models are guided to generate solutions with fewer steps without compromising accuracy. The method emphasizes training models on datasets combining complete and skipped reasoning paths, enabling them to learn efficient and accurate shortcuts.

The training framework comprises two main phases: initialization and iteration. The model is trained on a dataset containing comprehensive, step-by-step reasoning solutions during initialization. This establishes a foundational understanding of problem-solving. In the iteration phase, models are guided to generate shorter reasoning paths by reducing the number of steps in their responses. These shorter paths, verified for accuracy, are mixed with full-step solutions to create expanded datasets. Each iteration refines the model’s ability to identify and skip redundant steps, gradually improving efficiency. For instance, in tasks involving algebraic analogies, multi-digit arithmetic, and directional reasoning, the researchers generated datasets with detailed steps and selectively omitted certain steps to simulate human-like efficiency. These iterations allow the models to self-generate skipping data, refining their reasoning processes.

Empirical evaluations demonstrated the effectiveness of this approach across three tasks: algebraic analogies, multi-digit addition, and directional reasoning. Results highlighted that step-skipping enhanced both efficiency and generalization. For algebraic analogies, models achieved an accuracy increase of 4.76% in out-of-domain tasks, with a marked reduction in the number of reasoning steps. In multi-digit addition, performance improved by 13.91% in easier out-of-domain scenarios and by 4.75% in harder scenarios, underscoring the benefits of skipped reasoning steps. Similarly, directional reasoning tasks improved, with accuracy gains of up to 9.2% on challenging datasets. These results demonstrate that integrating skipped-step reasoning does not compromise task performance but enables models to solve problems more effectively and efficiently.

Further, the iterative training method showed that models could learn to balance accuracy and efficiency. Each iteration decreased the number of steps taken while maintaining or improving accuracy. By the fifth iteration, models consistently outperformed those trained solely on full-step datasets. This iterative refinement process also provided insights into the models’ ability to generalize to out-of-domain scenarios, suggesting that training on mixed datasets is instrumental in enhancing task-solving capabilities.

The study presents a significant advancement in equipping language models with human-like reasoning abilities. By incorporating step-skipping behavior, researchers demonstrated that models could achieve greater efficiency and maintain accuracy across diverse tasks. This approach addresses a critical limitation in existing models and opens avenues for future research on bridging the gap between human and machine reasoning. The contributions from leading institutions and companies underscore the collaborative efforts driving innovation in AI. The findings provide a promising direction for developing more efficient and versatile language models, paving the way for future advancements in artificial intelligence.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post This Research from Amazon Explores Step-Skipping Frameworks: Advancing Efficiency and Human-Like Reasoning in Language Models appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/26/this-research-from-amazon-explores-step-skipping-frameworks-advancing-efficiency-and-human-like-reasoning-in-language-models/feed/ 0 66734
This AI Paper Introduces G-NLL: A Novel Machine Learning Approach for Efficient and Accurate Uncertainty Estimation in Natural Language Generation https://www.marktechpost.com/2024/12/25/this-ai-paper-introduces-g-nll-a-novel-machine-learning-approach-for-efficient-and-accurate-uncertainty-estimation-in-natural-language-generation/ https://www.marktechpost.com/2024/12/25/this-ai-paper-introduces-g-nll-a-novel-machine-learning-approach-for-efficient-and-accurate-uncertainty-estimation-in-natural-language-generation/#respond Wed, 25 Dec 2024 20:48:58 +0000 https://www.marktechpost.com/?p=66695 Natural Language Generation (NLG) is a domain of artificial intelligence that seeks to enable machines to produce human-like text. By leveraging advancements in deep learning, researchers aim to develop systems capable of generating contextually relevant and coherent responses. Applications of this technology span diverse areas, including automated customer support, creative writing, and real-time language translation, […]

The post This AI Paper Introduces G-NLL: A Novel Machine Learning Approach for Efficient and Accurate Uncertainty Estimation in Natural Language Generation appeared first on MarkTechPost.

]]>

Natural Language Generation (NLG) is a domain of artificial intelligence that seeks to enable machines to produce human-like text. By leveraging advancements in deep learning, researchers aim to develop systems capable of generating contextually relevant and coherent responses. Applications of this technology span diverse areas, including automated customer support, creative writing, and real-time language translation, emphasizing seamless communication between humans and machines.

A key challenge in this domain lies in assessing the certainty of machine-generated text. Due to their probabilistic nature, language models may produce various outputs for the same input prompt. This variability raises concerns about the generated content’s reliability and the model’s confidence in its predictions. Addressing this issue is critical for applications where consistency and accuracy are paramount, such as medical or legal documentation.

To estimate uncertainty in generated text, traditional approaches rely on sampling multiple output sequences and analyzing them collectively. These methods, while insightful, demand significant computational resources since generating multiple sequences is computationally expensive. Consequently, the practicality of such methods diminishes for larger-scale deployments or tasks involving complex language models.

Researchers from the ELLIS Unit Linz and LIT AI Lab at Johannes Kepler University Linz, Austria, introduced a novel approach, G-NLL, to streamline the uncertainty estimation process. This method is based on computing the most probable output sequence’s negative log-likelihood (NLL). Unlike earlier approaches that rely on sampling, G-NLL uses greedy decoding to identify the most probable sequence and evaluate its likelihood. By focusing on this singular sequence, the method bypasses the need for extensive computational overhead, making it a more efficient alternative.

The G-NLL methodology involves calculating the probability of the most likely output sequence generated by a model. The negative log-likelihood of this sequence serves as a direct measure of uncertainty, with lower values indicating greater confidence in the generated text. This approach eliminates the redundancy of generating multiple sequences while maintaining the robustness required for effective uncertainty estimation. Further, the method integrates seamlessly with existing language models, requiring minimal modification to the decoding process.

Empirical evaluations of G-NLL demonstrated its superior performance across various tasks and models. Researchers tested the method on datasets commonly used for benchmarking language generation tasks, including machine translation and summarization. G-NLL consistently matched or surpassed the performance of traditional sampling-based methods. For instance, in a specific evaluation, the process reduced computational cost while maintaining accuracy levels on par with conventional techniques. Detailed results from experiments showed a significant efficiency improvement, with reduced computational demands by up to 50% in some tasks.

By addressing a critical limitation in NLG systems, the researchers provided a practical and scalable solution for estimating uncertainty. G-NLL represents a step forward in making language models more accessible for applications that require high reliability and computational efficiency. The innovation offers potential benefits for industries relying on automated text generation, including healthcare, education, and customer service, where confidence in outputs is crucial.

In conclusion, this research tackles the fundamental problem of uncertainty estimation in machine-generated text by introducing G-NLL. The method simplifies the process, reduces computational costs, and achieves strong performance across multiple benchmarks, solidifying its contribution to NLG. This advancement sets a new standard for efficiency and reliability in uncertainty estimation methods, paving the way for the broader adoption of language generation systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post This AI Paper Introduces G-NLL: A Novel Machine Learning Approach for Efficient and Accurate Uncertainty Estimation in Natural Language Generation appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/25/this-ai-paper-introduces-g-nll-a-novel-machine-learning-approach-for-efficient-and-accurate-uncertainty-estimation-in-natural-language-generation/feed/ 0 66695
This AI Paper Introduces ROMAS: A Role-Based Multi-Agent System for Efficient Database Monitoring and Planning https://www.marktechpost.com/2024/12/24/this-ai-paper-introduces-romas-a-role-based-multi-agent-system-for-efficient-database-monitoring-and-planning/ https://www.marktechpost.com/2024/12/24/this-ai-paper-introduces-romas-a-role-based-multi-agent-system-for-efficient-database-monitoring-and-planning/#respond Tue, 24 Dec 2024 20:39:31 +0000 https://www.marktechpost.com/?p=66675 Multi-agent systems (MAS) are pivotal in artificial intelligence, enabling multiple agents to work collaboratively to solve intricate tasks. These systems are designed to function in dynamic and unpredictable environments, addressing data analysis, process automation, and decision-making tasks. By incorporating advanced frameworks and leveraging large language models (LLMs), MAS has increased efficiency and adaptability for various […]

The post This AI Paper Introduces ROMAS: A Role-Based Multi-Agent System for Efficient Database Monitoring and Planning appeared first on MarkTechPost.

]]>

Multi-agent systems (MAS) are pivotal in artificial intelligence, enabling multiple agents to work collaboratively to solve intricate tasks. These systems are designed to function in dynamic and unpredictable environments, addressing data analysis, process automation, and decision-making tasks. By incorporating advanced frameworks and leveraging large language models (LLMs), MAS has increased efficiency and adaptability for various applications. However, enhancing their ability to handle real-world complexities remains a significant challenge.

A persistent issue in traditional MAS is their limited flexibility and adaptability. These systems often struggle with dynamic task requirements, relying on rigid task allocation and predefined unsuitable procedures for changing conditions. This rigidity increases the likelihood of errors and limits the system’s ability to recover effectively when deviations occur. Moreover, the lack of integrated mechanisms for self-planning and error correction exacerbates these inefficiencies, leading to wasted resources and suboptimal performance in complex scenarios.

Existing methods for MAS development include frameworks such as LangChain and AgentScope, which provide task allocation and development tools. While these frameworks facilitate the creation of agents and streamline deployment, they are limited by their inability to manage diverse data scenarios or provide robust solutions for advanced analytics. For example, traditional MAS systems like MetaGPT and AutoAgents lack global monitoring mechanisms and flexible agent generation, rendering them ineffective for tasks requiring dynamic adjustments and comprehensive error correction during execution.

Researchers from Ant Group and JD Group have introduced ROMAS, a Role-Based Multi-Agent System designed to address these limitations. ROMAS is built on the DB-GPT framework and incorporates role-based collaboration, enabling agents to take on specific roles such as planners, monitors, and workers. This innovative system facilitates real-time task monitoring, adaptive error correction, and low-code development. ROMAS enhances efficiency and scalability in database monitoring and planning tasks by supporting seamless deployment across various scenarios.

The ROMAS methodology emphasizes adaptability and robustness through its three operational phases: initialization, execution, and re-planning. In the initialization phase, the system divides tasks into subtasks and assigns them to specialized agents, each with distinct roles like data extraction, retrieval, and analysis. During execution, agents collaborate to complete tasks based on predefined strategies. A self-monitoring mechanism allows agents to identify and address errors dynamically, with unresolved issues escalated to a monitor for further analysis. The re-planning phase refines strategies using insights from the previous phases, ensuring alignment with the system’s objectives. The DB-GPT framework underpins ROMAS with powerful database handling, memory categorization, and self-reflection capabilities, allowing for effective task completion even in complex environments.

The researchers conducted extensive evaluations to demonstrate ROMAS’s performance, using datasets like FAMMA and HotpotQA to test its capabilities in domain-specific and general scenarios. On the FAMMA dataset, ROMAS achieved a success rate of 81.68%, while on the HotpotQA dataset, it reached 85.24%. These results highlight its superior performance to other MAS systems, including Generative Agents and AutoAgents. Marked features like the monitor mechanism and memory categorization contributed significantly to this success. The study also revealed that ROMAS reduced development complexity, with code volume decreasing to 1,500 rows compared to 2,500 rows in LangChain and 1,800 in AgentScope. Further, ROMAS demonstrated an average query processing time of 12.23 seconds, significantly faster than its counterparts.

Key findings include ROMAS’s ability to address pipeline and logical errors effectively. For instance, the system’s error correction mechanisms reduced error impact rates by 22.66% on average, showcasing its robust problem-solving capabilities. Integrating advanced memory mechanisms and the DB-GPT framework enhanced task efficiency by enabling seamless transitions between operational phases. These features improve system reliability and ensure that ROMAS maintains high adaptability across diverse scenarios.

In conclusion, ROMAS represents a significant advancement in multi-agent systems by addressing the critical limitations of traditional frameworks. Developed by Ant Group and JD Group researchers, the system leverages role-based collaboration, self-monitoring, and low-code deployment to streamline database monitoring and planning tasks. ROMAS has demonstrated superior performance through extensive evaluations, offering a scalable and efficient solution for complex analytical challenges. This innovation paves the way for further advancements in intelligent multi-agent systems and their applications.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post This AI Paper Introduces ROMAS: A Role-Based Multi-Agent System for Efficient Database Monitoring and Planning appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/24/this-ai-paper-introduces-romas-a-role-based-multi-agent-system-for-efficient-database-monitoring-and-planning/feed/ 0 66675
Evaluation Agent: A Multi-Agent AI Framework for Efficient, Dynamic, Multi-Round Evaluation, While Offering Detailed, User-Tailored Analyses https://www.marktechpost.com/2024/12/23/evaluation-agent-a-multi-agent-ai-framework-for-efficient-dynamic-multi-round-evaluation-while-offering-detailed-user-tailored-analyses/ https://www.marktechpost.com/2024/12/23/evaluation-agent-a-multi-agent-ai-framework-for-efficient-dynamic-multi-round-evaluation-while-offering-detailed-user-tailored-analyses/#respond Mon, 23 Dec 2024 19:20:21 +0000 https://www.marktechpost.com/?p=66644 Visual generative models have advanced significantly in terms of the ability to create high-quality images and videos. These developments, powered by AI, enable applications ranging from content creation to design. However, the capability of these models depends on the evaluation frameworks used to measure their performance, making efficient and accurate assessments a crucial area of […]

The post Evaluation Agent: A Multi-Agent AI Framework for Efficient, Dynamic, Multi-Round Evaluation, While Offering Detailed, User-Tailored Analyses appeared first on MarkTechPost.

]]>

Visual generative models have advanced significantly in terms of the ability to create high-quality images and videos. These developments, powered by AI, enable applications ranging from content creation to design. However, the capability of these models depends on the evaluation frameworks used to measure their performance, making efficient and accurate assessments a crucial area of focus.

Existing evaluation frameworks for visual generative models are often inefficient, requiring significant computational resources and rigid benchmarking processes. To measure performance, traditional tools rely heavily on large datasets and fixed metrics, such as FID and FVD. These methods lack flexibility and adaptability, often producing simple numerical scores without deeper interpretive insights. This creates a gap between the evaluation process and user-specific requirements, limiting their practicality in real-world applications.

Traditional benchmarks like VBench and EvalCrafter focus on specific dimensions such as subject consistency, aesthetic quality, and motion smoothness. However, these methods demand thousands of samples for evaluation, leading to high time costs. For instance, benchmarks like VBench require up to 4,355 samples per evaluation, consuming over 4,000 minutes of computation time. Despite their comprehensiveness, these frameworks struggle to adapt to user-defined criteria, leaving room for improvement in efficiency and flexibility.

Researchers from the Shanghai Artificial Intelligence Laboratory and Nanyang Technological University introduced the Evaluation Agent framework to address these limitations. This innovative solution mimics human-like strategies by conducting dynamic, multi-round evaluations tailored to user-defined criteria. Unlike rigid benchmarks, this approach integrates customizable evaluation tools, making it adaptable and efficient. The Evaluation Agent leverages large language models (LLMs) to power its intelligent planning and dynamic evaluation process.

The Evaluation Agent operates through two stages. The system identifies evaluation dimensions based on user input in the Proposal Stage and dynamically selects test cases. Prompts are generated by the PromptGen Agent, which designs tasks aligned with the user’s query. The Execution Stage involves generating visuals based on these prompts and evaluating them using an extensible toolkit. The framework eliminates redundant test cases and uncovers nuanced model behaviors by dynamically refining its focus. This dual-stage process allows for efficient evaluations while maintaining high accuracy.

The framework significantly outperforms traditional methods in terms of efficiency and adaptability. While benchmarks like VBench require thousands of samples and over 4,000 minutes to complete evaluations, the Evaluation Agent achieves similar accuracy using only 23 samples and 24 minutes per model dimension. Across various dimensions, such as aesthetic quality, spatial relationships, and motion smoothness, the Evaluation Agent demonstrated prediction accuracy comparable to established benchmarks while reducing computational costs by over 90%. For instance, the system evaluated models like VideoCrafter-2.0 with a consistency of up to 100% in multiple dimensions.

The Evaluation Agent achieved remarkable results in its experiments. It adapted to user-specific queries, providing detailed, interpretable results beyond numerical scores. It also supported evaluations across text-to-image (T2I) and text-to-video (T2V) models, highlighting its scalability and versatility. Considerable reductions in evaluation time were observed, from 563 minutes with T2I-CompBench to just 5 minutes for the same task using the Evaluation Agent. This efficiency positions the framework as a superior alternative for evaluating generative models in academic and industrial contexts.

The Evaluation Agent offers a transformative approach to visual generative model evaluation, overcoming the inefficiencies of traditional methods. By combining dynamic, human-like evaluation processes with advanced AI technologies, the framework provides a flexible and accurate solution for assessing diverse model capabilities. The substantial reduction in computational resources and time costs highlights its potential for broad adoption, paving the way for more effective evaluations in generative AI.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Evaluation Agent: A Multi-Agent AI Framework for Efficient, Dynamic, Multi-Round Evaluation, While Offering Detailed, User-Tailored Analyses appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/23/evaluation-agent-a-multi-agent-ai-framework-for-efficient-dynamic-multi-round-evaluation-while-offering-detailed-user-tailored-analyses/feed/ 0 66644
This AI Paper from aiXplain Introduces Bel Esprit: A Multi-Agent Framework for Building Accurate and Adaptive AI Model Pipelines https://www.marktechpost.com/2024/12/21/this-ai-paper-from-aixplain-introduces-bel-esprit-a-multi-agent-framework-for-building-accurate-and-adaptive-ai-model-pipelines/ https://www.marktechpost.com/2024/12/21/this-ai-paper-from-aixplain-introduces-bel-esprit-a-multi-agent-framework-for-building-accurate-and-adaptive-ai-model-pipelines/#respond Sat, 21 Dec 2024 20:34:00 +0000 https://www.marktechpost.com/?p=66591 Artificial intelligence has progressed from handling atomic tasks to addressing intricate, real-world problems requiring the integration of multiple specialized models. This approach, known as AI pipelines, allows for seamless task transitions by connecting different models to process diverse data inputs and outputs. These pipelines enable complex applications like multilingual video dubbing, multimodal content moderation, and […]

The post This AI Paper from aiXplain Introduces Bel Esprit: A Multi-Agent Framework for Building Accurate and Adaptive AI Model Pipelines appeared first on MarkTechPost.

]]>

Artificial intelligence has progressed from handling atomic tasks to addressing intricate, real-world problems requiring the integration of multiple specialized models. This approach, known as AI pipelines, allows for seamless task transitions by connecting different models to process diverse data inputs and outputs. These pipelines enable complex applications like multilingual video dubbing, multimodal content moderation, and advanced speech translation. The growing sophistication of AI pipelines reflects the increasing need for automated solutions that simplify and streamline challenging computational tasks in various domains.

Addressing complex computational challenges requires coordinating multiple models to handle different aspects of a problem. Current solutions often fall short when faced with ambiguous user requirements, poorly defined task parameters, and mismatched data modalities. For instance, computational tasks like multilingual dubbing demand careful alignment of inputs and outputs, such as matching audio transcription to translation models and text-to-speech synthesis. Such complexities make manual intervention necessary, slowing progress and leading to inefficiencies.

Existing methods for building AI pipelines often rely on static frameworks and predefined models tailored to specific tasks. While these approaches can handle isolated problems effectively, they lack adaptability. Manual adjustments are frequently required to address missing information, ensure semantic alignment, or resolve errors arising from mismatched modalities. Moreover, the rigidity of current systems limits their ability to cater to diverse user queries, leaving significant room for improvement in both flexibility and accuracy.

Researchers from aiXplain, Inc. and Los Gatos introduced a novel AI framework called Bel Esprit to overcome these challenges. This multi-agent system facilitates building customizable AI model pipelines tailored to user needs. Bel Esprit features specialized subagents, including Mentalist for clarifying user queries, Builder for pipeline assembly, and Inspector for error detection and correction. By employing a collaborative and iterative approach, the framework ensures pipelines are accurate and aligned with user intent. The system is designed to work dynamically, refining user inputs and optimizing the models chosen for specific tasks.

Bel Esprit is a graph-based framework with nodes representing AI functions and edges representing data flows. The Mentalist subagent begins by analyzing user queries to clarify ambiguous details, converting them into comprehensive task specifications. Builder then constructs an initial pipeline, breaking the task into manageable subgraphs. For example, distinct branches are created for each language in a multilingual dubbing task. The inspector reviews the pipeline for structural and semantic errors, ensuring alignment with the refined user requirements. This iterative process leverages techniques like chain-of-branches, where smaller subgraphs are built sequentially, facilitating model reuse and minimizing errors. Further, Bel Esprit integrates advanced large language models (LLMs) to automate reasoning and ensure seamless task execution.

The performance of Bel Esprit demonstrates its significant potential for transforming pipeline construction. The system achieved considerable results using exact match (EM) and graph edit distance (GED) metrics. The overall EM rate increased by 9.5%, indicating a higher rate of perfectly constructed pipelines. GED errors decreased by 28.1%, showcasing improvements in reducing discrepancies between generated and reference pipelines. For instance, when applied to multilingual video dubbing, Bel Esprit optimized workflows by reusing AI nodes, such as automatic speech recognition (ASR) models, across branches for different languages. This led to a streamlined pipeline construction process with fewer errors. Also, Bel Esprit effectively handled ambiguous user queries, with performance enhancements being more pronounced in cases where user input lacked clarity. The system’s iterative process ensured alignment with user intent, even in highly complex scenarios.

Bel Esprit significantly advances AI pipeline construction, addressing key ambiguity issues and error-prone assembly processes. Its innovative multi-agent collaboration, iterative refinement, and state-of-the-art models make it a robust solution for complex computational tasks. Bel Esprit sets a new benchmark for adaptability and precision in the field by automating critical stages of pipeline building and ensuring semantic accuracy. The framework’s demonstrated ability to improve efficiency and handle complex queries underscores its potential as a transformative tool in AI applications.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post This AI Paper from aiXplain Introduces Bel Esprit: A Multi-Agent Framework for Building Accurate and Adaptive AI Model Pipelines appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/21/this-ai-paper-from-aixplain-introduces-bel-esprit-a-multi-agent-framework-for-building-accurate-and-adaptive-ai-model-pipelines/feed/ 0 66591
Google DeepMind Introduces FACTS Grounding: A New AI Benchmark for Evaluating Factuality in Long-Form LLM Response https://www.marktechpost.com/2024/12/20/google-deepmind-introduces-facts-grounding-a-new-ai-benchmark-for-evaluating-factuality-in-long-form-llm-response/ https://www.marktechpost.com/2024/12/20/google-deepmind-introduces-facts-grounding-a-new-ai-benchmark-for-evaluating-factuality-in-long-form-llm-response/#respond Sat, 21 Dec 2024 00:11:20 +0000 https://www.marktechpost.com/?p=66567 Despite the transformative potential of large language models (LLMs), these models face significant challenges in generating contextually accurate responses faithful to the provided input. Ensuring factuality in LLM outputs is particularly critical in tasks requiring responses grounded in lengthy, complex documents, which form the basis for advancing their applications in research, education, and industry. One […]

The post Google DeepMind Introduces FACTS Grounding: A New AI Benchmark for Evaluating Factuality in Long-Form LLM Response appeared first on MarkTechPost.

]]>

Despite the transformative potential of large language models (LLMs), these models face significant challenges in generating contextually accurate responses faithful to the provided input. Ensuring factuality in LLM outputs is particularly critical in tasks requiring responses grounded in lengthy, complex documents, which form the basis for advancing their applications in research, education, and industry.

One major challenge in LLM development is their tendency to produce inaccurate or “hallucinated” content. This issue arises when models generate plausible-sounding text that is not supported by the input data. Such inaccuracies can have severe consequences, including the spread of misinformation and decreased trust in AI systems. Addressing this problem requires comprehensive benchmarks that evaluate the fidelity of LLM outputs to ensure that the generated text aligns strictly with the context provided in a prompt.

Existing solutions to factuality challenges involve supervised fine-tuning and reinforcement learning. These methods aim to optimize LLMs to adhere more closely to factual content, albeit with limitations. Another approach leverages inference-time strategies like advanced prompting and model state interpretability to reduce inaccuracies. However, these techniques often result in trade-offs, compromising qualities such as creativity and response diversity. Consequently, there remains a need for a robust and scalable framework to systematically evaluate and enhance LLMs’ factuality without sacrificing other attributes.

Researchers from Google DeepMind, Google Research, Google Cloud, and Kaggle introduced the FACTS Grounding Leaderboard to address these gaps. This benchmark is specifically designed to measure LLMs’ ability to generate responses fully grounded in extensive input contexts. The dataset includes user requests paired with source documents of up to 32,000 tokens, demanding responses that are factually correct and adhere strictly to the input context. The leaderboard is hosted on Kaggle and includes public and private data splits, encouraging broad participation while maintaining dataset integrity.

The methodology underlying the FACTS Grounding benchmark involves a two-stage evaluation process. First, responses are screened for eligibility, disqualifying those failing to address user requests adequately. Eligible responses are then evaluated for factuality using multiple automated judge models, including Gemini 1.5 Pro, GPT-4o, and Claude 3.5 Sonnet. These models are prompted with optimized templates, ensuring high alignment with human judgment. For instance, the evaluation process uses span-level analysis to validate each claim in the response, with scores aggregated across multiple models to minimize bias. Further, the benchmark incorporates measures to prevent gaming of the scoring system, such as requiring comprehensive responses that directly address user queries.

The FACTS Grounding Leaderboard revealed diverse performance results across tested models, showcasing the benchmark’s rigor in evaluating factuality. Among the models evaluated, Gemini 1.5 Flash achieved an impressive factuality score of 85.8% in the public dataset, while Gemini 1.5 Pro and GPT-4o followed closely with scores of 84.9% and 83.6%, respectively. On the private dataset, Gemini 1.5 Pro outperformed others with a score of 90.7%. The disqualification of ineligible responses reduced scores by 1% to 5%, emphasizing the importance of robust filtering mechanisms. These results highlight the benchmark’s ability to differentiate performance and promote transparency in model evaluation.

The FACTS Grounding Leaderboard fills a critical gap in evaluating LLMs by focusing on long-form response generation. Unlike benchmarks emphasizing narrow use cases, such as short-form factuality or summarization, this benchmark addresses a broader spectrum of tasks, including fact-finding, document analysis, and information synthesis. By maintaining high evaluation standards and actively updating the leaderboard with new models, the initiative provides an essential tool for advancing the factual accuracy of LLMs.

The research team’s efforts underscore the importance of rigorous evaluation frameworks in overcoming the challenges associated with LLM-generated content. The FACTS Grounding benchmark provides a systematic approach to measuring factuality and fosters innovation in developing more reliable and accurate AI systems. This work sets a new standard for evaluating LLMs and inspires further advancements in artificial intelligence.


Check out the Paper and Technical Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Google DeepMind Introduces FACTS Grounding: A New AI Benchmark for Evaluating Factuality in Long-Form LLM Response appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/20/google-deepmind-introduces-facts-grounding-a-new-ai-benchmark-for-evaluating-factuality-in-long-form-llm-response/feed/ 0 66567
Microsoft AI Research Open-Sources PromptWizard: A Feedback-Driven AI Framework for Efficient and Scalable LLM Prompt Optimization https://www.marktechpost.com/2024/12/18/microsoft-ai-research-open-sources-promptwizard-a-feedback-driven-ai-framework-for-efficient-and-scalable-llm-prompt-optimization/ https://www.marktechpost.com/2024/12/18/microsoft-ai-research-open-sources-promptwizard-a-feedback-driven-ai-framework-for-efficient-and-scalable-llm-prompt-optimization/#respond Wed, 18 Dec 2024 23:26:31 +0000 https://www.marktechpost.com/?p=66499 One of the crucial factors in achieving high-quality outputs from these models lies in the design of prompts—carefully crafted input instructions that guide the model to produce the desired responses. Despite their importance, prompt creation is a labor-intensive process that often requires domain-specific knowledge and significant human effort. These limitations have spurred the development of […]

The post Microsoft AI Research Open-Sources PromptWizard: A Feedback-Driven AI Framework for Efficient and Scalable LLM Prompt Optimization appeared first on MarkTechPost.

]]>

One of the crucial factors in achieving high-quality outputs from these models lies in the design of prompts—carefully crafted input instructions that guide the model to produce the desired responses. Despite their importance, prompt creation is a labor-intensive process that often requires domain-specific knowledge and significant human effort. These limitations have spurred the development of automated systems to refine and optimize prompts efficiently.

One of the significant challenges in prompt engineering is the reliance on manual expertise to tailor prompts for each unique task. This approach is time-consuming and needs to scale more effectively for complex or domain-specific applications. Furthermore, existing methods for optimizing prompts are often restricted to open-source models that provide access to internal computations. Black-box systems, such as proprietary models accessible only via APIs, present an additional hurdle, as their internal workings are opaque, making traditional gradient-based techniques impractical. These constraints highlight the urgent need for solutions that work efficiently with limited resources while remaining effective across diverse tasks.

Currently, methods for prompt optimization can be broadly classified into two categories: continuous and discrete approaches. Continuous techniques, such as soft prompts, rely on auxiliary models to refine instructions but require substantial computational resources and are not directly applicable to black-box systems. Discrete methods, including approaches like PromptBreeder and EvoPrompt, focus on generating variations of prompts and selecting the best-performing ones based on evaluation metrics. While these approaches have shown promise, they often need more structured feedback mechanisms to improve. They need to balance exploration with task-specific refinements, leading to suboptimal results.

Researchers from Microsoft Research India have developed and open-sourced PromptWizard, an innovative AI framework for optimizing prompts in black-box LLMs. This framework employs a feedback-driven critique-and-synthesis mechanism to iteratively refine prompt instructions and in-context examples iteratively, enhancing task performance. PromptWizard stands out by combining guided exploration with structured critiques to ensure the holistic improvement of prompts. Unlike earlier methods, it aligns task-specific requirements with a systematic optimization process, offering an efficient and scalable solution for diverse NLP applications.

PromptWizard operates through two primary phases: a generation phase and a test-time inference phase. During the generation phase, the system uses LLMs to create multiple variations of a base prompt by applying cognitive heuristics. These variations are evaluated against training examples to identify high-performing candidates. The framework integrates a critique mechanism that analyzes the strengths and weaknesses of each prompt, generating feedback that informs subsequent iterations of refinement. By synthesizing new examples and leveraging reasoning chains, the system enhances both the diversity and quality of prompts. The optimized prompts and examples are applied to unseen tasks at test time, ensuring consistent performance improvements. This approach significantly reduces computational overhead by focusing on meaningful refinements rather than random mutations, making it suitable for resource-constrained environments.

The framework’s effectiveness is demonstrated through extensive experiments across 45 tasks, including datasets like Big Bench Instruction Induction (BBII) and arithmetic reasoning benchmarks such as GSM8K, AQUARAT, and SVAMP. PromptWizard achieved the highest accuracy in zero-shot settings on 13 out of 19 tasks, outperforming baseline methods like Instinct and EvoPrompt. It further improved accuracy in one-shot scenarios, leading to 16 out of 19 tasks. For example, it achieved a zero-shot accuracy of 90% on GSM8K and 82.3% on SVAMP, showcasing its ability to handle complex reasoning tasks effectively. Further, PromptWizard reduced token usage and API calls by up to 60 times compared to discrete methods like PromptBreeder, with a total cost of only $0.05 per task, making it one of the most cost-efficient solutions available.

PromptWizard’s success lies in its innovative combination of sequential optimization, guided critiques, and expert persona integration, ensuring task-specific alignment and interpretability. The results highlight its potential to transform prompt engineering, offering a scalable, efficient, and accessible solution for optimizing LLMs across diverse domains. This advancement underscores the importance of integrating automated frameworks into NLP workflows, paving the way for more effective and affordable utilization of advanced AI technologies.


Check out the Paper, Blog, and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Microsoft AI Research Open-Sources PromptWizard: A Feedback-Driven AI Framework for Efficient and Scalable LLM Prompt Optimization appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/18/microsoft-ai-research-open-sources-promptwizard-a-feedback-driven-ai-framework-for-efficient-and-scalable-llm-prompt-optimization/feed/ 0 66499
Researchers from Sakana AI Introduce NAMMs: Optimized Memory Management for Efficient and High-Performance Transformer Models https://www.marktechpost.com/2024/12/17/researchers-from-sakana-ai-introduce-namms-optimized-memory-management-for-efficient-and-high-performance-transformer-models/ https://www.marktechpost.com/2024/12/17/researchers-from-sakana-ai-introduce-namms-optimized-memory-management-for-efficient-and-high-performance-transformer-models/#respond Tue, 17 Dec 2024 16:39:43 +0000 https://www.marktechpost.com/?p=66449 Transformers have become the backbone of deep learning models for tasks requiring sequential data processing, such as natural language understanding, computer vision, and reinforcement learning. These models rely heavily on self-attention mechanisms, enabling them to capture complex relationships within input sequences. However, as tasks and models scale, the demand for longer context windows increases significantly. […]

The post Researchers from Sakana AI Introduce NAMMs: Optimized Memory Management for Efficient and High-Performance Transformer Models appeared first on MarkTechPost.

]]>

Transformers have become the backbone of deep learning models for tasks requiring sequential data processing, such as natural language understanding, computer vision, and reinforcement learning. These models rely heavily on self-attention mechanisms, enabling them to capture complex relationships within input sequences. However, as tasks and models scale, the demand for longer context windows increases significantly. Managing this extended context window efficiently is crucial because it impacts performance and computational cost. Despite their strength, transformers face challenges in maintaining efficiency while handling long-context inputs, making this an active area of research.

One of the significant challenges is balancing performance with resource efficiency. Transformers store previously computed representations in a memory cache known as the Key-Value (KV) cache, allowing them to reference past inputs efficiently. However, this KV cache grows exponentially for long-context tasks, consuming substantial memory and computational resources. Existing approaches attempt to reduce the KV cache size by removing less important tokens, but these methods rely on manually designed heuristics. The limitations of these approaches are evident: they often lead to performance degradation, as token removal strategies are not optimized to retain essential information for downstream tasks.

Current tools, such as H2O and L2 methods, attempt to alleviate this problem by introducing metrics like L2 norms and entropy to quantify token importance. These approaches aim to selectively prune tokens from the KV cache, reducing memory usage while preserving model performance. Despite some success, these methods introduce an inherent trade-off—reducing the memory footprint results in a performance loss. Models using these techniques struggle to generalize across tasks, and their heuristic-driven design prevents significant improvements in both performance and efficiency simultaneously.

A research team from Sakana AI, Japan, has introduced Neural Attention Memory Models (NAMMs). NAMMs are a new class of memory management models that dynamically optimize the KV cache in transformers. Instead of relying on hand-designed rules, NAMMs learn token importance through evolutionary optimization. By conditioning on the attention matrices of transformers, NAMMs enable each layer to retain only the most relevant tokens, enhancing both efficiency and performance without altering the base transformer architecture. This universality makes NAMMs applicable to any transformer-based model, as their design depends solely on features extracted from attention matrices.

The methodology behind NAMMs involves extracting meaningful features from the attention matrix using a spectrogram-based technique. The researchers apply the Short-Time Fourier Transform (STFT) to compress the attention values into a spectrogram representation. This compact representation captures how token importance evolves across the attention span. The spectrogram features are then reduced using an exponential moving average (EMA) operation to minimize complexity. NAMMs use a lightweight neural network to evaluate these compressed features and assign a selection score to each token. Tokens with low selection scores are evicted from the KV cache, freeing up memory while ensuring performance is not compromised.

A critical innovation in NAMMs is the introduction of backward attention mechanisms. This design allows the network to compare tokens efficiently, preserving only the most relevant occurrences while discarding redundant ones. By leveraging cross-token communication, NAMMs optimize memory usage dynamically across layers, ensuring transformers retain crucial long-range information for each task.

The performance of NAMMs was rigorously evaluated across multiple benchmarks, showcasing their superiority over existing methods. On the LongBench benchmark, NAMMs improved normalized performance by 11% while reducing the KV cache size to 25% of the original model. Similarly, on the challenging InfiniteBench benchmark, where average input lengths exceed 200,000 tokens, NAMMs outperformed baseline models by increasing performance from 1.05% to 11%. This result highlights NAMMs’ ability to scale effectively for long-context tasks without sacrificing accuracy. Moreover, the memory footprint of NAMMs on InfiniteBench was reduced to approximately 40% of the original size, demonstrating their efficiency in managing long sequences.

The researchers further validated NAMMs’ versatility through zero-shot transfer experiments. NAMMs trained exclusively on natural language tasks were applied to new transformers and input modalities, including computer vision and reinforcement learning models. For instance, when tested with a Llava Next Video 7B model on long video understanding tasks, NAMMs improved the base model’s performance while maintaining a reduced memory footprint. In reinforcement learning experiments using Decision Transformers on continuous control tasks, NAMMs achieved an average performance gain of 9% across multiple tasks, demonstrating their ability to discard unhelpful information and improve decision-making capabilities.

In conclusion, NAMMs provide a powerful solution to the challenge of long-context processing in transformers. By learning efficient memory management strategies through evolutionary optimization, NAMMs overcome the limitations of hand-designed heuristics. The results demonstrate that transformers equipped with NAMMs achieve superior performance while significantly reducing computational costs. Their universal applicability and success across diverse tasks highlight their potential to advance transformer-based models across multiple domains, marking a significant step toward efficient long-context modeling.


Check out the Paper and Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Researchers from Sakana AI Introduce NAMMs: Optimized Memory Management for Efficient and High-Performance Transformer Models appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/17/researchers-from-sakana-ai-introduce-namms-optimized-memory-management-for-efficient-and-high-performance-transformer-models/feed/ 0 66449
This AI Paper from Microsoft and Novartis Introduces Chimera: A Machine Learning Framework for Accurate and Scalable Retrosynthesis Prediction https://www.marktechpost.com/2024/12/16/this-ai-paper-from-microsoft-and-novartis-introduces-chimera-a-machine-learning-framework-for-accurate-and-scalable-retrosynthesis-prediction/ https://www.marktechpost.com/2024/12/16/this-ai-paper-from-microsoft-and-novartis-introduces-chimera-a-machine-learning-framework-for-accurate-and-scalable-retrosynthesis-prediction/#respond Tue, 17 Dec 2024 07:50:27 +0000 https://www.marktechpost.com/?p=66446 Chemical synthesis is essential in developing new molecules for medical applications, materials science, and fine chemicals. This process, which involves planning chemical reactions to create desired target molecules, has traditionally relied on human expertise. Recent advancements have turned to computational methods to enhance the efficiency of retrosynthesis—working backward from a target molecule to determine the […]

The post This AI Paper from Microsoft and Novartis Introduces Chimera: A Machine Learning Framework for Accurate and Scalable Retrosynthesis Prediction appeared first on MarkTechPost.

]]>

Chemical synthesis is essential in developing new molecules for medical applications, materials science, and fine chemicals. This process, which involves planning chemical reactions to create desired target molecules, has traditionally relied on human expertise. Recent advancements have turned to computational methods to enhance the efficiency of retrosynthesis—working backward from a target molecule to determine the series of reactions needed to synthesize it. By leveraging modern computational techniques, researchers aim to solve long-standing bottlenecks in synthetic chemistry, making these processes faster and more accurate.

One of the critical challenges in retrosynthesis is accurately predicting chemical reactions that are rare or less frequently encountered. These reactions, although uncommon, are vital for designing novel chemical pathways. Traditional machine-learning models often fail to predict these reactions due to insufficient representation in training data. Also, multi-step retrosynthesis planning errors can cascade, leading to invalid synthetic routes. This limitation hinders the ability to explore innovative and diverse pathways for chemical synthesis, particularly in cases requiring uncommon reactions.

Existing computational methods for retrosynthesis have primarily focused on single-step models or rule-based expert systems. These methods rely on pre-defined rules or extensive training datasets, which limits their adaptability to new and unique reaction types. For instance, some approaches use graph-based or sequence-based models to predict the most likely transformations. While these methods have improved accuracy for common reactions, they often need more flexibility to account for the complexities and nuances of rare chemical transformations, leading to a gap in comprehensive retrosynthetic planning.

Researchers from Microsoft Research, Novartis Biomedical Research, and Jagiellonian University developed Chimera, an ensemble framework for retrosynthesis prediction. Chimera integrates outputs from multiple machine-learning models with diverse inductive biases, combining their strengths through a learned ranking mechanism. This approach leverages two newly developed state-of-the-art models: NeuralLoc, which focuses on molecule editing using graph neural networks, and R-SMILES 2, a de-novo model employing a sequence-to-sequence Transformer architecture. By combining these models, Chimera enhances both accuracy and scalability for retrosynthetic predictions.

The methodology behind Chimera relies on combining outputs from its constituent models through a ranking system that assigns scores based on model agreement and predictive confidence. NeuralLoc encodes molecular structures as graphs, enabling precise prediction of reaction sites and templates. This method ensures that predicted transformations align closely with known chemical rules while maintaining computational efficiency. Meanwhile, R-SMILES 2 utilizes advanced attention mechanisms, including Group-Query Attention, to predict reaction pathways. This model’s architecture also incorporates improvements in normalization and activation functions, ensuring superior gradient flow and inference speed. Chimera combines these predictions, using overlap-based scoring to rank potential pathways. This integration ensures that the framework balances the strengths of editing-based and de-novo approaches, enabling robust predictions even for complex and rare reactions.

The performance of Chimera has been rigorously validated against publicly available datasets such as USPTO-50K and USPTO-FULL, as well as the proprietary Pistachio dataset. On USPTO-50K, Chimera achieved a 1.7% improvement in top-10 prediction accuracy over the previous state-of-the-art methods, demonstrating its capability to accurately predict both common and rare reactions. On USPTO-FULL, it further improved top-10 accuracy by 1.6%. Scaling the model to the Pistachio dataset, which contains over three times the data of USPTO-FULL, showed that Chimera maintained high accuracy across a broader range of reactions. Expert comparisons with organic chemists revealed that Chimera’s predictions were consistently preferred over individual models, confirming its effectiveness in practical applications.

The framework was also tested on an internal Novartis dataset of over 10,000 reactions to evaluate its robustness under distribution shifts. In this zero-shot setting, where no additional fine-tuning was performed, Chimera demonstrated superior accuracy compared to its constituent models. This highlights its capability to generalize across datasets and predict viable synthetic pathways even in real-world scenarios. Further, Chimera excelled in multi-step retrosynthesis tasks, achieving close to 100% success rates on benchmarks such as SimpRetro, significantly outperforming individual models. The framework’s ability to find pathways for highly challenging molecules further underscores its potential to transform computational retrosynthesis.

Chimera represents a groundbreaking advancement in retrosynthesis prediction by addressing the challenges of rare reaction prediction and multi-step planning. The framework demonstrates superior accuracy and scalability by integrating diverse models and employing a robust ranking mechanism. With its ability to generalize across datasets and excel in complex retrosynthetic tasks, Chimera is set to accelerate progress in chemical synthesis, paving the way for innovative approaches to molecular design.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post This AI Paper from Microsoft and Novartis Introduces Chimera: A Machine Learning Framework for Accurate and Scalable Retrosynthesis Prediction appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/16/this-ai-paper-from-microsoft-and-novartis-introduces-chimera-a-machine-learning-framework-for-accurate-and-scalable-retrosynthesis-prediction/feed/ 0 66446