- 替换 Google Translate 为智谱 GLM-4-Flash API - LLM 总结论文/项目核心内容生成中文标题 - 标题简洁有力,突出技术亮点 - 简化输出格式:标题 + 链接(无摘要) - 添加翻译进度显示
68 lines
2.6 KiB
Markdown
68 lines
2.6 KiB
Markdown
# AI Daily Brief - 2026-02-28
|
|
|
|
> Collected at: 2/28/2026, 12:08:50 AM
|
|
> Total items: 131
|
|
|
|
## 🔥 Top 10 Highlights
|
|
|
|
1. sponsors/muratcankoylan
|
|
https://github.com/sponsors/muratcankoylan
|
|
|
|
2. login?return_to=%2Fruvnet%2Fclaude-flow
|
|
https://github.com/login?return_to=%2Fruvnet%2Fclaude-flow
|
|
|
|
3. Search More, Think Less: Rethinking Long-Horizon Agentic Search for Efficiency and Generalization
|
|
https://huggingface.co/papers/2602.22675
|
|
|
|
4. AgentDropoutV2: Optimizing Information Flow in Multi-Agent Systems via Test-Time Rectify-or-Reject Pruning
|
|
https://huggingface.co/papers/2602.23258
|
|
|
|
5. Accelerating Diffusion via Hybrid Data-Pipeline Parallelism Based on Conditional Guidance Scheduling
|
|
https://huggingface.co/papers/2602.21760
|
|
|
|
6. Exploratory Memory-Augmented LLM Agent via Hybrid On- and Off-Policy Optimization
|
|
https://huggingface.co/papers/2602.23008
|
|
|
|
7. login?return_to=%2Fruvnet%2Fwifi-densepose
|
|
https://github.com/login?return_to=%2Fruvnet%2Fwifi-densepose
|
|
|
|
8. login?return_to=%2Fbytedance%2Fdeer-flow
|
|
https://github.com/login?return_to=%2Fbytedance%2Fdeer-flow
|
|
|
|
9. login?return_to=%2Fmoonshine-ai%2Fmoonshine
|
|
https://github.com/login?return_to=%2Fmoonshine-ai%2Fmoonshine
|
|
|
|
10. sponsors/obra
|
|
https://github.com/sponsors/obra
|
|
|
|
## 📂 Categories
|
|
|
|
### Agent Frameworks
|
|
|
|
- Toward Expert Investment Teams:A Multi-Agent LLM System with Fine-Grained Trading Tasks
|
|
http://arxiv.org/abs/2602.23330v1
|
|
- AgentDropoutV2: Optimizing Information Flow in Multi-Agent Systems via Test-Time Rectify-or-Reject Pruning
|
|
http://arxiv.org/abs/2602.23258v1
|
|
|
|
### AI Infrastructure / Inference Optimization
|
|
|
|
- Bitwise Systolic Array Architecture for Runtime-Reconfigurable Multi-precision Quantized Multiplication on Hardware Accelerators
|
|
http://arxiv.org/abs/2602.23334v1
|
|
- Invariant Transformation and Resampling based Epistemic-Uncertainty Reduction
|
|
http://arxiv.org/abs/2602.23315v1
|
|
- Agency and Architectural Limits: Why Optimization-Based Systems Cannot Be Norm-Responsive
|
|
http://arxiv.org/abs/2602.23239v1
|
|
- InnerQ: Hardware-aware Tuning-free Quantization of KV Cache for Large Language Models
|
|
http://arxiv.org/abs/2602.23200v1
|
|
- Assessing Deanonymization Risks with Stylometry-Assisted LLM Agent
|
|
http://arxiv.org/abs/2602.23079v1
|
|
- Rejection Mixing: Fast Semantic Propagation of Mask Tokens for Efficient DLLM Inference
|
|
http://arxiv.org/abs/2602.22868v1
|
|
- Differentiable Zero-One Loss via Hypersimplex Projections
|
|
http://arxiv.org/abs/2602.23336v1
|
|
- FairQuant: Fairness-Aware Mixed-Precision Quantization for Medical Image Classification
|
|
http://arxiv.org/abs/2602.23192v1
|
|
|
|
---
|
|
*Generated by AINewsCollector*
|