Author: Mohammad Asjad

Mohammad Asjad
231 POSTS0 COMMENTS
Asjad is an intern consultant at Marktechpost. He is persuing B.Tech in mechanical engineering at the Indian Institute of Technology, Kharagpur. Asjad is a Machine learning and deep learning enthusiast who is always researching the applications of machine learning in healthcare.

Researchers from Snowflake and CMU Introduce SuffixDecoding: A Novel Model-Free Approach to Accelerating Large Language Model (LLM) Inference through Speculative Decoding

Large language models (LLMs) have rapidly become a foundational component of today's consumer and enterprise applications. However, the need for a fast generation of...

The Semantic Hub: A Cognitive Approach to Language Model Representations

Language models have demonstrated remarkable capabilities in processing diverse data types, including multilingual text, code, mathematical expressions, images, and audio. However, a fundamental question...

Researchers from Stanford and Cornell Introduce APRICOT: A Novel AI Approach that Merges LLM-based Bayesian Active Preference Learning with Constraint-Aware Task Planning

In the rapidly evolving field of household robotics, a significant challenge has emerged in executing personalized organizational tasks, such as arranging groceries in a...

Nearest Neighbor Normalization: A Sublinear Approach to Improving Contrastive Retrieval

Contrastive image and text models face significant challenges in optimizing retrieval accuracy despite their crucial role in large-scale text-to-image and image-to-text retrieval systems. While...

Predicting and Interpreting In-Context Learning Curves Through Bayesian Scaling Laws

Large Language Models (LLMs) have demonstrated remarkable in-context learning (ICL) capabilities, where they can learn tasks from demonstrations without requiring additional training. A critical...

Multi-Scale Geometric Analysis of Language Model Features: From Atomic Patterns to Galaxy Structures

Large Language Models (LLMs) have emerged as powerful tools in natural language processing, yet understanding their internal representations remains a significant challenge. Recent breakthroughs...

AUTO-CEI: A Curriculum and Expert Iteration Approach to Elevate LLMs’ Response Precision and Control Refusal Rates Across Diverse Reasoning Domains

Large language models (LLMs) are increasingly utilized for complex reasoning tasks, requiring them to provide accurate responses across various challenging scenarios. These tasks include...

CodeFavor: A Machine Learning Framework that Trains Pairwise Preference Models with Synthetic Code Preferences Generated from Code Evolution like Code Commits and Code Critiques

Large Language Models (LLMs) have revolutionized software development by enabling code completion, functional code generation from instructions, and complex code modifications for bug fixes...

SimpleToM: Evaluating Applied Theory of Mind Capabilities in Large Language Models

Theory of Mind (ToM) capabilities - the ability to attribute mental states and predict behaviors of others - have become increasingly critical as Large...

LongRAG: A Robust RAG Framework for Long-Context Question Answering

Large Language Models (LLMs) have revolutionized long-context question answering (LCQA), a complex task requiring reasoning over extensive documents to provide accurate answers. While recent...

MiniCTX: Advancing Context-Dependent Theorem Proving in Large Language Models

Formal theorem proving has emerged as a critical benchmark for assessing the reasoning capabilities of large language models (LLMs), with significant implications for mathematical...

Meta AI Researchers Introduce Token-Level Detective Reward Model (TLDR) to Provide Fine-Grained Annotations for Large Vision Language Models

Vision Language Models (VLMs) have demonstrated remarkable capabilities in generating human-like text in response to images, with notable examples including GPT-4, Gemini, PaLiGemma, LLaVA,...