[FEEDBACK] Daily Papers

Note that this is not a post about adding new papers, it's about feedback on the Daily Papers community update feature.
How to submit a paper to the Daily Papers, like @akhaliq (AK)?
- Submitting is available to paper authors
- Only recent papers (less than 7d) can be featured on the Daily
Then drop the arxiv id in the form at https://huggingface.co/papers/submit
- Add medias to the paper (images, videos) when relevant
- You can start the discussion to engage with the community
Please check out the documentation
We are excited to share our recent work on MLLM architecture design titled "Ovis: Structural Embedding Alignment for Multimodal Large Language Model".
Paper: https://arxiv.org/abs/2405.20797
Github: https://github.com/AIDC-AI/Ovis
Model: https://huggingface.co/AIDC-AI/Ovis-Clip-Llama3-8B
Data: https://huggingface.co/datasets/AIDC-AI/Ovis-dataset
@Yiwen-ntu for now we support only videos as paper covers in the Daily.
we are excited to share our work titled "Hierarchical Prompting Taxonomy: A Universal Evaluation Framework for Large Language Models" : https://arxiv.org/abs/2406.12644
Dear AK and HF Team,
Iβm super excited to recommend SE Arena, a new interactive platform for benchmarking Software Engineering chatbots.
π If youβre working with AI in software dev or just passionate about improving how these models perform in real-world dev workflows, you have to check it out!
π§ The best part? SE Arena has a transparent, open-source leaderboard, and you can actively contribute by casting your votes to shape the evaluations. Plus, with RepoChat, it pulls in real repo context (issues, commits, PRs) to make things feel real.
π£ Want to get involved and help drive the future of AI in software engineering? Head over to https://huggingface.co/spaces/SE-Arena/Software-Engineering-Arena and cast your vote today! π
Our paper is published in FORGE 2025: https://conf.researchr.org/details/forge-2025/forge-2025-papers/6/SE-Arena-An-Interactive-Platform-for-Evaluating-Foundation-Models-in-Software-Engine
Check the details in https://arxiv.org/abs/2502.01860
Weβd love your feedback and contributions! π
Can we align LLMs with personal preferences? It is hard to collect individual annotations sufficiently and train LLMs for each persona.... The answer is
Yes! ποΈ Drift achieves personalized alignment only with 50~100 examples.
- Drift Approximation: For efficient preference modeling, we first define various attributes and find the best composite of them to explain given examples.
- Differential prompting: We don't need to construct attribute-dedicated datasets! We show differential prompting to evaluate each attribute in a zero-shot manner.
- Drift Decoding: We can align LLM with the composite of attributes in a training-free manner! We don't need expensive LLM training and savings for each user.
We prove theoretically each objective of the approximation and decoding stages, and for all stages, there is no gradient computation in the total process.
Check the details here! https://arxiv.org/abs/2502.14289
ππDeepSQL-R1-distill-8B : A Quantized DeepSeek AI Model for SQL Code Generation
π₯ Outperforms Llama-3.2, Mistral-7B, and Claude-3 Sonnet in SQL generation tasks.
β‘ Superior execution accuracy and faster inference speeds for complex SQL queries.
π Optimized for efficiency with quantization & distillation techniques.
Model Link: https://huggingface.co/imsanjoykb/deepSQL-R1-distill-8B
Code Link: https://github.com/imsanjoykb/deepSQL-R1-distill-8B
Paper : https://doi.org/10.6084/m9.figshare.28330301.v1
Inference: https://drive.google.com/file/d/145PP-oW50OMS1bYJaYuUphfufpsuOGWl/view?usp=sharing
Hello AK and HF Team,
We would to add our recent paper "Exploring Intrinsic Normal Prototypes within a Single Image for Universal Anomaly Detection" in HF daily papers. I am putting this request here, since I don't already have a paper in HF daily papers.
Paper: https://arxiv.org/pdf/2503.02424
Github:https://github.com/luow23/INP-Former
Thanks,
Wei Luo
Hi,
I came across this paper for synthetic data generation. "TabularARGN: A Flexible and Efficient Auto-Regressive Framework for Generating High-Fidelity Synthetic Data"
Paper: https://arxiv.org/html/2501.12012v1
GitHub: https://github.com/mostly-ai/mostlyai
Regards,
David
Hello AK and HF Team,
We would like to add our paper "GenConViT: Deepfake Video Detection Using Generative Convolutional Vision Transformer"
paper: https://arxiv.org/abs/2307.07036
code: https://github.com/erprogs/GenConViT
Thank you!,
Deressa
doi:10.57967/hf/4791
π ScrapeGoat Music Models: Now Available! π΅
After our recent work on HCF-Aware Training to optimize memory efficiency in language models (especially beneficial for specialized models like ours), we're excited to announce that all training and weights for the ScrapeGoat Music project are now available in our Hugging Face repositories:
ALL training python codes and weights available in the repo on HF.
Dear AK and HF Team,
Exciting Update! We're thrilled to share WritingBench: A Comprehensive Framework for Evaluating Generative Writing
π [Paper] β’ π [Github Repo] β’ π [Critic Model] β’ βοΈ [Writing Model]
π‘WritingBench is a comprehensive benchmark for evaluating LLMs' writing capabilities across 1,239 real-world queries, spanning:
- 6 primary domains and 100 fine-grained subdomains
- 3 core writing requirements: Style / Format / Length
- 1,546 avg. tokens per query, integrating diverse sources of materials
- Each query is paired with 5 instance-specific criteria
Regards,
Yuning