PulseAugur
LIVE 16:59:32
research · [11 sources] ·
0
research

New AI frameworks tackle concept extraction, taxonomy generation, and materials discovery

Researchers have developed SC-Taxo, a framework using large language models (LLMs) to generate semantically consistent hierarchical taxonomies for scientific literature. This approach addresses inconsistencies in existing methods by employing a bidirectional generation mechanism that refines both bottom-up abstractions and top-down constraints. Experiments show SC-Taxo improves hierarchy alignment and heading quality, even generalizing to Chinese scientific literature. AI

Summary written by gemini-2.5-flash-lite from 11 sources. How we write summaries →

IMPACT Improves organization and retrieval of scientific knowledge, potentially accelerating research and discovery.

RANK_REASON This is a research paper detailing a new framework for taxonomy generation.

Read on arXiv cs.LG →

COVERAGE [11]

  1. arXiv cs.CL TIER_1 · Shiqiang Cai, Nianhong Niu, Shizhu He, Kang Liu, Jun Zhao ·

    SC-Taxo: Hierarchical Taxonomy Generation under Semantic Consistency Constraints using Large Language Models

    arXiv:2605.00620v1 Announce Type: new Abstract: Scientific literature is expanding at an unprecedented pace, making it increasingly challenging to efficiently organize and access domain knowledge. A high-quality scientific taxonomy offers a structured and hierarchical representat…

  2. arXiv cs.CL TIER_1 · Jun Zhao ·

    SC-Taxo: Hierarchical Taxonomy Generation under Semantic Consistency Constraints using Large Language Models

    Scientific literature is expanding at an unprecedented pace, making it increasingly challenging to efficiently organize and access domain knowledge. A high-quality scientific taxonomy offers a structured and hierarchical representation of a research field, facilitating literature…

  3. arXiv cs.AI TIER_1 · Usha Bhalla, Thomas Fel, Can Rager, Sheridan Feucht, Tal Haklay, Daniel Wurgaft, Siddharth Boppana, Matthew Kowal, Vasudev Shyam, Jack Merullo, Atticus Geiger, Ekdeep Singh Lubana ·

    Do Sparse Autoencoders Capture Concept Manifolds?

    arXiv:2604.28119v1 Announce Type: cross Abstract: Sparse autoencoders (SAEs) are widely used to extract interpretable features from neural network representations, often under the implicit assumption that concepts correspond to independent linear directions. However, a growing bo…

  4. arXiv cs.AI TIER_1 · Ekdeep Singh Lubana ·

    Do Sparse Autoencoders Capture Concept Manifolds?

    Sparse autoencoders (SAEs) are widely used to extract interpretable features from neural network representations, often under the implicit assumption that concepts correspond to independent linear directions. However, a growing body of evidence suggests that many concepts are ins…

  5. Hugging Face Daily Papers TIER_1 ·

    Do Sparse Autoencoders Capture Concept Manifolds?

    Sparse autoencoders (SAEs) are widely used to extract interpretable features from neural network representations, often under the implicit assumption that concepts correspond to independent linear directions. However, a growing body of evidence suggests that many concepts are ins…

  6. arXiv cs.CL TIER_1 · Yunze Jia, Yuehui Xian, Yangyang Xu, Pengfei Dang, Xiangdong Ding, Jun Sun, Yumei Zhou, Dezhen Xue ·

    Semantic Embeddings of Chemical Elements for Enhanced Materials Inference and Discovery

    arXiv:2502.14912v2 Announce Type: replace Abstract: We present a framework for generating universal semantic embeddings of chemical elements to advance materials inference and discovery. This framework leverages ElementBERT, a domain-specific BERT-based natural language processin…

  7. arXiv cs.CL TIER_1 · Tomer Ashuach, Dana Arad, Aaron Mueller, Martin Tutek, Yonatan Belinkov ·

    CRISP: Persistent Concept Unlearning via Sparse Autoencoders

    arXiv:2508.13650v3 Announce Type: replace Abstract: As large language models (LLMs) are increasingly deployed in real-world applications, the need to selectively remove unwanted knowledge while preserving model utility has become paramount. Recent work has explored sparse autoenc…

  8. arXiv cs.LG TIER_1 · Mingze Li, Yu Rong, Songyou Li, Lihong Wang, Jiacheng Cen, Liming Wu, Anyi Li, Zongzhao Li, Qiuliang Liu, Rui Jiao, Tian Bian, Pengju Wang, Hao Sun, Jianfeng Zhang, Ji-Rong Wen, Deli Zhao, Shifeng Jin, Tingyang Xu, Wenbing Huang ·

    Agentic Fusion of Large Atomic and Language Models to Accelerate Materials Discovery

    arXiv:2604.23758v1 Announce Type: new Abstract: The discovery of novel materials is critical for global energy and quantum technology transitions. While deep learning has fundamentally reshaped this landscape, existing predictive or generative models typically operate in isolatio…

  9. Hugging Face Daily Papers TIER_1 ·

    A Unifying Framework for Unsupervised Concept Extraction

    Techniques for concept extraction, such as sparse autoencoders and transcoders, aim to extract high-level symbolic concepts from low-level nonsymbolic representations. When these extracted concepts are used for downstream tasks such as model steering and unlearning, it is essenti…

  10. arXiv stat.ML TIER_1 · Chandler Squires, Pradeep Ravikumar ·

    A Unifying Framework for Unsupervised Concept Extraction

    arXiv:2604.24936v1 Announce Type: cross Abstract: Techniques for concept extraction, such as sparse autoencoders and transcoders, aim to extract high-level symbolic concepts from low-level nonsymbolic representations. When these extracted concepts are used for downstream tasks su…

  11. arXiv stat.ML TIER_1 · Pradeep Ravikumar ·

    A Unifying Framework for Unsupervised Concept Extraction

    Techniques for concept extraction, such as sparse autoencoders and transcoders, aim to extract high-level symbolic concepts from low-level nonsymbolic representations. When these extracted concepts are used for downstream tasks such as model steering and unlearning, it is essenti…