computer-vision

Tag

Cards List
#computer-vision

@elonmusk: The human-perceived RGB is image 1 and the Tesla AI photon count reconstruction is image 2. This is why Tesla FSD can s…

X AI KOLs Following · 10h ago Cached

Elon Musk explains that Tesla FSD utilizes AI photon count reconstruction rather than standard RGB, enabling superior performance in low-light and high-glare conditions.

0 favorites 0 likes
#computer-vision

@elonmusk: Tesla AI Vision

X AI KOLs Following · 12h ago

A brief mention of Tesla AI Vision, referring to Tesla's computer vision-based approach to autonomous driving.

0 favorites 0 likes
#computer-vision

@HuggingPapers: Microsoft just released Phi-Ground-Any on Hugging Face A 4B parameter vision model for GUI grounding that achieves SOTA…

X AI KOLs Following · 22h ago Cached

Microsoft has released Phi-Ground-Any, a 4B parameter vision model for GUI grounding on Hugging Face that achieves state-of-the-art results, enabling AI agents to precisely interact with screen elements.

0 favorites 0 likes
#computer-vision

@tenderizzation: it’s literally off the scale! welcome back yolov3

X AI KOLs Following · yesterday Cached

A social media post expresses excitement about the return or renewed relevance of the YOLOv3 object detection model.

0 favorites 0 likes
#computer-vision

@Tesla: Tesla Vision allows us to deploy airbags up to 70 milliseconds earlier if your Tesla detects an unavoidable collision T…

X AI KOLs Following · yesterday Cached

Tesla announces its Vision system can detect unavoidable collisions and deploy airbags up to 70 milliseconds earlier, potentially making the difference between serious injury and walking away from a crash.

0 favorites 0 likes
#computer-vision

@neil_xbt: Andrej Karpathy could have charged $1,000 for this computer vision lecture! He put it on YouTube. The man who built Tes…

X AI KOLs Timeline · yesterday

Andrej Karpathy released a free computer vision lecture on YouTube covering image captioning, localization, segmentation and transfer learning from his production experience at Tesla and OpenAI.

0 favorites 1 likes
#computer-vision

FoodCHA: Multi-Modal LLM Agent for Fine-Grained Food Analysis

arXiv cs.AI · 2d ago Cached

This paper introduces FoodCHA, a multi-modal LLM agent framework designed for fine-grained food analysis, addressing challenges in hierarchical consistency and attribute discrimination for dietary monitoring.

0 favorites 0 likes
#computer-vision

Intelligent CCTV for Urban Design: AI-Based Analysis of Soft Infrastructure at Intersections

arXiv cs.AI · 2d ago Cached

This academic paper introduces an AI-enabled analytics framework using existing CCTV infrastructure to evaluate the impact of soft traffic interventions on vehicle speed and safety at urban intersections.

0 favorites 0 likes
#computer-vision

HNC: Leveraging Hard Negative Captions towards Models with Fine-Grained Visual-Linguistic Comprehension Capabilities

arXiv cs.CL · 2d ago Cached

The paper introduces Hard Negative Captions (HNC), a dataset and method for training vision-language models to achieve fine-grained comprehension by addressing weak associations in web-collected image-text pairs.

0 favorites 0 likes
#computer-vision

UK Cars to Get AI Cameras to Detect Impaired Drivers

Reddit r/ArtificialInteligence · 2d ago Cached

The UK is implementing AI-powered camera systems in vehicles to detect impaired drivers, enhancing road safety through real-time monitoring.

0 favorites 0 likes
#computer-vision

Visual Perceptual to Conceptual First-Order Rule Learning Networks [R]

Reddit r/MachineLearning · 3d ago Cached

This paper introduces gammaILP, a fully differentiable framework for learning first-order rules directly from image data without label leakage, addressing challenges in symbol grounding and predicate invention.

0 favorites 0 likes
#computer-vision

SwiftI2V: Efficient High-Resolution Image-to-Video Generation via Conditional Segment-wise Generation

Hugging Face Daily Papers · 3d ago Cached

SwiftI2V is a new efficient framework for high-resolution image-to-video generation that uses conditional segment-wise generation to achieve 2K synthesis with significantly reduced computational costs. It enables practical generation on single consumer or datacenter GPUs while maintaining input fidelity.

0 favorites 0 likes
#computer-vision

Continuous-Time Distribution Matching for Few-Step Diffusion Distillation

Hugging Face Daily Papers · 3d ago Cached

This paper introduces Continuous-Time Distribution Matching (CDM), a method for few-step diffusion distillation that migrates from discrete to continuous optimization to improve visual fidelity and preserve fine details.

0 favorites 0 likes
#computer-vision

@lillyguisnet: WEEE!!! I had not had the opportunity to try SAM3.1 yet, but simply prompting for "worm" perfectly segmented my images!…

X AI KOLs Following · 3d ago Cached

A user shares enthusiastic feedback about SAM 3.1's ability to accurately segment images using simple text prompts like 'worm', highlighting significant improvements over SAM 1.

0 favorites 0 likes
#computer-vision

StableI2I: Spotting Unintended Changes in Image-to-Image Transition

Hugging Face Daily Papers · 4d ago Cached

This paper introduces StableI2I, a reference-free evaluation framework for assessing content fidelity and consistency in image-to-image generation tasks. It also presents StableI2I-Bench, a benchmark for evaluating multi-modal language models on these assessment tasks.

0 favorites 0 likes
#computer-vision

Lightning Unified Video Editing via In-Context Sparse Attention

Hugging Face Daily Papers · 4d ago Cached

This paper introduces In-context Sparse Attention (ISA), a framework that significantly reduces computational costs in video editing by pruning redundant context and using dynamic query grouping. The authors demonstrate the method's effectiveness with LIVEditor, achieving near-lossless acceleration and state-of-the-art results on multiple video editing benchmarks.

0 favorites 0 likes
#computer-vision

TT4D: A Pipeline and Dataset for Table Tennis 4D Reconstruction From Monocular Videos

Hugging Face Daily Papers · 2026-05-02 Cached

This paper introduces TT4D, a novel pipeline and large-scale dataset for reconstructing table tennis gameplay in 4D from monocular videos. It features a unique lift-first approach that estimates 3D ball trajectories and spin before time segmentation, enabling robust reconstruction even with occlusions.

0 favorites 0 likes
#computer-vision

MoCapAnything V2: End-to-End Motion Capture for Arbitrary Skeletons

Papers with Code Trending · 2026-04-30 Cached

MoCapAnything V2 introduces a fully end-to-end framework for arbitrary-skeleton motion capture from monocular video, jointly optimizing video-to-pose and pose-to-rotation predictions to resolve rotation ambiguity.

0 favorites 0 likes
#computer-vision

Representation Fréchet Loss for Visual Generation

Papers with Code Trending · 2026-04-30 Cached

This paper introduces FD-loss, a method to optimize Fréchet Distance as a training objective for visual generation by decoupling population and batch sizes. It demonstrates that this approach improves generator quality and suggests FID may not always accurately reflect visual quality.

0 favorites 0 likes
#computer-vision

Solving the “Whac-a-mole dilemma”: A smarter way to debias AI vision models

MIT News — Artificial Intelligence · 2026-04-29 Cached

Researchers from MIT, WPI, and Google propose WRING, a novel post-processing debiasing method for Vision-Language Models that avoids the 'Whac-a-mole dilemma' of amplifying other biases when removing specific ones.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback