-
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails
Paper • 2502.05163 • Published • 23 -
CRANE: Reasoning with constrained LLM generation
Paper • 2502.09061 • Published • 21 -
Investigating the Impact of Quantization Methods on the Safety and Reliability of Large Language Models
Paper • 2502.15799 • Published • 7 -
AISafetyLab: A Comprehensive Framework for AI Safety Evaluation and Improvement
Paper • 2502.16776 • Published • 6
Collections
Discover the best community collections!
Collections including paper arxiv:2502.17125
-
Language Agent Tree Search Unifies Reasoning Acting and Planning in Language Models
Paper • 2310.04406 • Published • 10 -
Tree of Thoughts: Deliberate Problem Solving with Large Language Models
Paper • 2305.10601 • Published • 14 -
Language Models as Compilers: Simulating Pseudocode Execution Improves Algorithmic Reasoning in Language Models
Paper • 2404.02575 • Published • 50 -
Voyager: An Open-Ended Embodied Agent with Large Language Models
Paper • 2305.16291 • Published • 11
-
Chain-of-Verification Reduces Hallucination in Large Language Models
Paper • 2309.11495 • Published • 39 -
Hallucination Detox: Sensitive Neuron Dropout (SeND) for Large Language Model Training
Paper • 2410.15460 • Published • 1 -
DeCoRe: Decoding by Contrasting Retrieval Heads to Mitigate Hallucinations
Paper • 2410.18860 • Published • 11 -
Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models
Paper • 2411.14257 • Published • 14
-
KRLabsOrg/lettucedect-base-modernbert-en-v1
Token Classification • 0.1B • Updated • 1.76k • 17 -
KRLabsOrg/lettucedect-large-modernbert-en-v1
Token Classification • 0.4B • Updated • 310 • 28 -
LettuceDetect: A Hallucination Detection Framework for RAG Applications
Paper • 2502.17125 • Published • 12 -
LettuceDetect
🥬7Let Us Detect your hallucinations! Demo for our framework.
-
LoRA+: Efficient Low Rank Adaptation of Large Models
Paper • 2402.12354 • Published • 7 -
The FinBen: An Holistic Financial Benchmark for Large Language Models
Paper • 2402.12659 • Published • 23 -
TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization
Paper • 2402.13249 • Published • 14 -
TrustLLM: Trustworthiness in Large Language Models
Paper • 2401.05561 • Published • 69
-
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails
Paper • 2502.05163 • Published • 23 -
CRANE: Reasoning with constrained LLM generation
Paper • 2502.09061 • Published • 21 -
Investigating the Impact of Quantization Methods on the Safety and Reliability of Large Language Models
Paper • 2502.15799 • Published • 7 -
AISafetyLab: A Comprehensive Framework for AI Safety Evaluation and Improvement
Paper • 2502.16776 • Published • 6
-
KRLabsOrg/lettucedect-base-modernbert-en-v1
Token Classification • 0.1B • Updated • 1.76k • 17 -
KRLabsOrg/lettucedect-large-modernbert-en-v1
Token Classification • 0.4B • Updated • 310 • 28 -
LettuceDetect: A Hallucination Detection Framework for RAG Applications
Paper • 2502.17125 • Published • 12 -
LettuceDetect
🥬7Let Us Detect your hallucinations! Demo for our framework.
-
Language Agent Tree Search Unifies Reasoning Acting and Planning in Language Models
Paper • 2310.04406 • Published • 10 -
Tree of Thoughts: Deliberate Problem Solving with Large Language Models
Paper • 2305.10601 • Published • 14 -
Language Models as Compilers: Simulating Pseudocode Execution Improves Algorithmic Reasoning in Language Models
Paper • 2404.02575 • Published • 50 -
Voyager: An Open-Ended Embodied Agent with Large Language Models
Paper • 2305.16291 • Published • 11
-
LoRA+: Efficient Low Rank Adaptation of Large Models
Paper • 2402.12354 • Published • 7 -
The FinBen: An Holistic Financial Benchmark for Large Language Models
Paper • 2402.12659 • Published • 23 -
TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization
Paper • 2402.13249 • Published • 14 -
TrustLLM: Trustworthiness in Large Language Models
Paper • 2401.05561 • Published • 69
-
Chain-of-Verification Reduces Hallucination in Large Language Models
Paper • 2309.11495 • Published • 39 -
Hallucination Detox: Sensitive Neuron Dropout (SeND) for Large Language Model Training
Paper • 2410.15460 • Published • 1 -
DeCoRe: Decoding by Contrasting Retrieval Heads to Mitigate Hallucinations
Paper • 2410.18860 • Published • 11 -
Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models
Paper • 2411.14257 • Published • 14