hypedarhypedar
feedtrendsdiscovershowcasearchive
login
login
login
FeedTrendsDiscoverShowcaseArchiveDashboard
Submit Showcase

Trending now

Workflow + Code Generation + Automation62Policy + Ethics53Dev Tools + Agents + Automation51
View all trends →

hypedar

AI trend radar for developers. Catch emerging papers, repos, and discussions before the hype peaks.

AboutGitHubDiscord

By the makers of hypedar

Codepawl

Open-source tools for developers.

Explore our tools →
AboutPrivacyTermsX

© 2026 Codepawl

Built by Codepawl·© 2026

About·Terms·Privacy·Security

GitHub·Discord·X

feedtrendsdiscovershowcasearchive
← feed
arXiv13h ago
4.8

Cards Against LLMs: Benchmarking Humor Alignment in Large Language Models

Yousra Fettach, Guillaume Bied, Hannu Toivonen, Tijl De Bie

View original ↗

Analysis

Viral velocity
low
Implementation gapYES
Novelty6/10
Categorypaper
Topics
reasoningalignmentevaluation

Opportunity Brief

Build a benchmark suite that evaluates LLMs on subjective human cultural alignment tasks like humor. Developers should create an extensible framework to ingest varied datasets (e.g., jokes, puns, sarcasm) to measure 'humor-alignment' across different models.

Suggested repo: nanoEval-Humor

"Is your model actually funny or just regurgitating datasets?"

Estimated effort: 20h