Author: Sajjad Ansari

Sajjad Ansari
107 POSTS0 COMMENTS
Sajjad Ansari is a final year undergraduate from IIT Kharagpur. As a Tech enthusiast, he delves into the practical applications of AI with a focus on understanding the impact of AI technologies and their real-world implications. He aims to articulate complex AI concepts in a clear and accessible manner.

SPARE: Training-Free Representation Engineering for Managing Knowledge Conflicts in Large Language Models

Large Language Models (LLMs) have demonstrated impressive capabilities in handling knowledge-intensive tasks through their parametric knowledge stored within model parameters. However, the stored knowledge...

WorFBench: A Benchmark for Evaluating Complex Workflow Generation in Large Language Model Agents

Large Language Models (LLMs) have shown remarkable potential in solving complex real-world problems, from function calls to embodied planning and code generation. A critical...

UC Berkeley Researchers Propose DocETL: A Declarative System that Optimizes Complex Document Processing Tasks using LLMs

Large Language Models (LLMs) have gained significant attention in data management, with applications spanning data integration, database tuning, query optimization, and data cleaning. However,...

DIFFUSEARCH: Revolutionizing Chess AI with Implicit Search and Discrete Diffusion Modeling

Large Language Models (LLMs) have gained significant attention in AI research due to their impressive capabilities. However,  their limitation lies with long-term planning and...

Self-Data Distilled Fine-Tuning: A Solution for Pruning and Supervised Fine-tuning Challenges in LLMs

Large language models (LLMs) like GPT-4, Gemini, and Llama 3 have revolutionized natural language processing through extensive pre-training and supervised fine-tuning (SFT). However, these...

Meissonic: A Non-Autoregressive Mask Image Modeling Text-to-Image Synthesis Model that can Generate High-Resolution Images

Large Language Models (LLMs) have demonstrated remarkable progress in natural language processing tasks, inspiring researchers to explore similar approaches for text-to-image synthesis. At the...

Google DeepMind Research Introduces Diversity-Rewarded CFG Distillation: A Novel Finetuning Approach to Enhance the Quality-Diversity Trade-off in Generative AI Models

Generative AI models, driven by Large Language Models (LLMs) or diffusion techniques, are revolutionizing creative domains like art and entertainment. These models can generate...

OPTIMA: Enhancing Efficiency and Effectiveness in LLM-Based Multi-Agent Systems

Large Language Models (LLMs) have gained significant attention for their versatility in various tasks, from natural language processing to complex reasoning. A promising application...

CausalMM: A Causal Inference Framework that Applies Structural Causal Modeling to Multimodal Large Language Models (MLLMs)

Multimodal Large Language Models (MLLMs) have made significant progress in various applications using the power of Transformer models and their attention mechanisms. However, these...

Researchers from UCSD and Adobe Introduce Presto!: An AI Approach to Inference Acceleration for Score-based Diffusion Transformers via Reducing both Sampling Steps and Cost...

Text-to-Audio (TTA) and Text-to-Music (TTM) generation have seen significant advancements in recent years, driven by audio-domain diffusion models. These models have demonstrated superior audio...

Apple Researchers Propose BayesCNS: A Unified Bayesian Approach Tackling Cold Start and Non-Stationarity in Large-Scale Search Systems

Information Retrieval (IR) systems for search and recommendations often utilize Learning-to-Rank (LTR) solutions to prioritize relevant items for user queries. These models heavily depend...

Enhancing Text Retrieval: Overcoming the Limitations with Contextual Document Embeddings

Text retrieval in machine learning faces significant challenges in developing effective methods for indexing and retrieving documents. Traditional approaches relied on sparse lexical matching...