|
Caching image prototype embeddings for image-guided object detection using OWL-ViT
|
|
2
|
482
|
January 30, 2026
|
|
[Quiestion]How to specify 'model_type' of 'Qwen/Qwen3-VL-8B-Instruct-GGUF'?
|
|
3
|
10
|
January 30, 2026
|
|
SAM3Video: CLIPTextModelOutput passed as tensor causes crash with text prompts
|
|
0
|
10
|
January 29, 2026
|
|
Different lm_head size and vocab_size
|
|
1
|
908
|
January 28, 2026
|
|
Custom KV Cache Steering Implementation Fails with IndexError in LLaVA Generation
|
|
1
|
12
|
January 28, 2026
|
|
Transformers v5 timelines
|
|
1
|
34
|
January 28, 2026
|
|
Issue: Discrepancy Between Layer-Wise Density Plots vs. Mean Trajectory Plots in LLaVA-1.5 Attention Analysis
|
|
2
|
18
|
January 25, 2026
|
|
[Discussion] Validating Attention Map Visualization for Visual Fading in LLaVA-1.5
|
|
4
|
37
|
January 23, 2026
|
|
No fix for High Vulnerabilities in transformers latest package
|
|
2
|
31
|
January 22, 2026
|
|
How to disable caching in .from_pretrained()
|
|
4
|
1228
|
January 18, 2026
|
|
DetLLM – Deterministic Inference Checks
|
|
0
|
22
|
January 17, 2026
|
|
Distributed LLaMA Inference Engine Built from Scratch (KV Cache, GQA, RoPE)
|
|
0
|
26
|
January 16, 2026
|
|
Run name issue, different run name file in webpage & local
|
|
1
|
88
|
January 16, 2026
|
|
Whisper fine-tuned with custom tokens works with model.generate but doesn't with a pipeline()
|
|
3
|
38
|
January 14, 2026
|
|
GPT 2 finetuning peaks at 8 GiB of VRAM
|
|
7
|
78
|
January 12, 2026
|
|
Model_accepts_loss_kwargs detection based on **kwargs is too permissive
|
|
2
|
260
|
January 5, 2026
|
|
Seeking Advice🔥🔥| Strategy for Embedding Multiple Subjective Reviews in One-time Event Domain Recommendations
|
|
2
|
42
|
January 23, 2026
|
|
TurboTensors: Optimizing CPU LLM Performance
|
|
0
|
22
|
December 31, 2025
|
|
Significant generation degradation and repetition loops when enabling KV-cache for Qwen3-VL
|
|
2
|
89
|
December 29, 2025
|
|
Injecting multi modal embeddings into a language model breaks the `generate` function
|
|
1
|
88
|
December 28, 2025
|
|
Transformers v4 or v5 for my new project?
|
|
1
|
70
|
December 27, 2025
|
|
Assistant model is not passed onto the custom_generate method
|
|
3
|
24
|
December 25, 2025
|
|
How can i get TRANSFORMERS_CACHE in transformers v5?
|
|
2
|
49
|
December 19, 2025
|
|
CDM-CTM Fusion: A Rigorous Framework for Depth-Aware Autoregressive Control
|
|
0
|
19
|
December 13, 2025
|
|
Tensor Dimension Mismatch when using TRL GKDTrainer
|
|
3
|
20
|
December 12, 2025
|
|
Transformers.js need for token to char mapping
|
|
3
|
35
|
December 11, 2025
|
|
[Pipelines] Mask Generation Parameters
|
|
2
|
121
|
December 10, 2025
|
|
Having trouble to configure trainer for T5 model evaluation
|
|
1
|
42
|
December 9, 2025
|
|
How do I speedup my callbacks and reduce stall before they start?
|
|
1
|
38
|
December 9, 2025
|
|
Getting 429 Too Many Request
|
|
3
|
135
|
December 8, 2025
|