Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
Chirag Agarwal's picture
5 10

Chirag Agarwal

AikyamLab
·
https://chirag-agarwall.github.io/
  • _cagarwal
  • chirag-agarwal-0a6a43a1

AI & ML interests

Explainability and Interpretability; AI Safety; AI Alignment

Organizations

LLM-XAI's profile picture Aikyam Lab's profile picture

upvoted a paper 2 months ago

CURE-Med: Curriculum-Informed Reinforcement Learning for Multilingual Medical Reasoning

Paper • 2601.13262 • Published Jan 19 • 2
upvoted 2 papers 4 months ago

CLINIC: Evaluating Multilingual Trustworthiness in Language Models for Healthcare

Paper • 2512.11437 • Published Dec 12, 2025 • 4

Polarity-Aware Probing for Quantifying Latent Alignment in Language Models

Paper • 2511.21737 • Published Nov 21, 2025 • 1
upvoted a collection 4 months ago

Polarity-Aware Probing Datasets

Collection
Datasets for PA-Probing described in "Polarity-Aware Probing for Quantifying Latent Alignment in Language Models" https://www.arxiv.org/pdf/2511.21737 • 2 items • Updated Dec 7, 2025 • 1
upvoted a paper about 2 years ago

Faithfulness vs. Plausibility: On the (Un)Reliability of Explanations from Large Language Models

Paper • 2402.04614 • Published Feb 7, 2024 • 3
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs