deep-learning

Tag

Cards List
#deep-learning

@mubeitech: The Transformer is not the endgame of AI, says NVIDIA VP of AI Research Sanja Fidler.

X AI KOLs Timeline · 2026-04-20 Cached

Sanja Fidler, VP of AI Research at NVIDIA and head of the company’s spatial-intelligence lab, says the Transformer’s Achilles heel is clear: training costs are sky-high and the hunger for data is bottomless. A new architectural breakthrough is overdue, and next-gen variants are already emerging.

0 favorites 0 likes
#deep-learning

Deep Learning-Based Amharic Chatbot for University FAQs

arXiv cs.CL · 2026-04-20 Cached

This paper presents a deep learning-based chatbot system for answering frequently asked questions in the Amharic language at universities, achieving 91.55% accuracy using neural networks with TensorFlow and Keras. The system addresses Amharic-specific linguistic challenges including morphological variation and lexical gaps, and was deployed on Facebook Messenger via Heroku.

0 favorites 0 likes
#deep-learning

Seeing the Intangible: Survey of Image Classification into High-Level and Abstract Categories

arXiv cs.CL · 2026-04-20 Cached

A comprehensive survey examining image classification into high-level and abstract categories, clarifying the tacit understanding of high-level semantics in computer vision through multidisciplinary analysis of commonsense, emotional, aesthetic, and interpretative semantics. The paper identifies persistent challenges in abstract concept image classification and emphasizes the importance of hybrid AI systems for addressing complex visual reasoning tasks.

0 favorites 0 likes
#deep-learning

@techyoutbe: 1.5-Hour Lecture on “LLM Architecture” by Stanford

X AI KOLs Timeline · 2026-04-19 Cached

Stanford University offers a 1.5-hour lecture on LLM architecture covering fundamental concepts and design principles of large language models.

0 favorites 0 likes
#deep-learning

TwinTrack: Post-hoc Multi-Rater Calibration for Medical Image Segmentation

Hugging Face Daily Papers · 2026-04-17 Cached

TwinTrack is a post-hoc calibration framework for pancreatic cancer segmentation that aligns ensemble model probabilities with the empirical mean human response across multiple annotators, improving interpretability and calibration metrics on multi-rater benchmarks.

0 favorites 0 likes
#deep-learning

ArtifactNet: Detecting AI-Generated Music via Forensic Residual Physics

Hugging Face Daily Papers · 2026-04-17 Cached

ArtifactNet is a lightweight neural network framework that detects AI-generated music by analyzing codec-specific artifacts in audio signals, achieving F1=0.9829 on a new 6,183-track benchmark (ArtifactBench) with 49x fewer parameters than competing methods. The approach uses forensic physics principles to extract codec residuals through a bounded-mask UNet and compact CNN, with codec-aware training reducing cross-codec drift by 83%.

0 favorites 0 likes
#deep-learning

NTIRE 2026 Challenge on Video Saliency Prediction: Methods and Results

Hugging Face Daily Papers · 2026-04-16 Cached

This paper presents the NTIRE 2026 Challenge on Video Saliency Prediction, introducing a novel dataset of 2,000 diverse videos with saliency maps collected via crowdsourced mouse tracking from over 5,000 assessors. Over 20 teams participated, with 7 passing the final phase, and all data is made publicly available.

0 favorites 0 likes
#deep-learning

huggingface/transformers Patch release v5.5.4

GitHub Releases Watchlist · 2026-04-13 Cached

Hugging Face releases transformers library patch version v5.5.4, a routine maintenance update to the widely-used NLP/deep learning framework.

0 favorites 0 likes
#deep-learning

Augmenting citizen science with computer vision for fish monitoring

MIT News — Artificial Intelligence · 2026-03-25 Cached

Researchers from MIT and the Woodwell Climate Research Center published a paper on using computer vision to automate fish monitoring, improving upon traditional citizen science methods for river herring conservation.

0 favorites 0 likes
#deep-learning

Can AI help predict which heart-failure patients will worsen within a year?

MIT News — Artificial Intelligence · 2026-03-12 Cached

MIT researchers have developed PULSE-HF, a deep learning model that predicts whether heart failure patients will experience worsening left ventricular ejection fraction within a year using electrocardiograms. The model, published in Lancet eClinical Medicine, could help clinicians prioritize high-risk patients and reduce unnecessary hospital visits in both well-resourced and low-resource clinical settings.

0 favorites 0 likes
#deep-learning

Using AI to perceive the universe in greater depth

Google DeepMind Blog · 2025-10-24 Cached

DeepMind introduces Deep Loop Shaping, a novel AI method that reduces noise and improves feedback control in gravitational wave observatories, reducing noise by 30-100x in LIGO's most unstable feedback loops and enabling detection of hundreds more astronomical events annually.

0 favorites 0 likes
#deep-learning

Discovering new solutions to century-old problems in fluid dynamics

Google DeepMind Blog · 2025-10-24 Cached

DeepMind researchers discovered new families of unstable singularities in fundamental fluid dynamics equations using AI techniques, potentially advancing understanding of century-old mathematical problems like the Navier-Stokes equations. The work collaborates with Brown, NYU, and Stanford, revealing patterns in blow-up behavior with unprecedented computational accuracy.

0 favorites 0 likes
#deep-learning

UI-TARS-2 Technical Report: Advancing GUI Agent with Multi-Turn Reinforcement Learning

Papers with Code Trending · 2025-09-02 Cached

UI-TARS-2 is a native GUI-centered agent model that addresses data scalability, multi-turn RL, and environment stability challenges, achieving state-of-the-art results on GUI benchmarks (88.2 on Online-Mind2Web, 47.5 on OSWorld, 50.6 on WindowsAgentArena,73.3 on AndroidWorld) and outperforming Claude and OpenAI agents.

0 favorites 0 likes
#deep-learning

Elon Musk wanted an OpenAI for-profit

OpenAI Blog · 2024-12-13 Cached

OpenAI publishes an article discussing Elon Musk's vision for an OpenAI for-profit structure, emphasizing that hardware capabilities and computational scaling are fundamental drivers of AI breakthroughs, with predictions about near-term progress in robotics, theorem-proving, and AI competitiveness.

0 favorites 0 likes
#deep-learning

Simplifying, stabilizing, and scaling continuous-time consistency models

OpenAI Blog · 2024-10-23 Cached

OpenAI presents sCM (simplified continuous-time consistency models), a new approach that scales consistency models to 1.5B parameters and achieves ~50x speedup over diffusion models by generating high-quality samples in just 2 steps. The method demonstrates comparable sample quality to state-of-the-art diffusion models while using less than 10% of the effective sampling compute.

0 favorites 0 likes
#deep-learning

MuZero: Checkmate For Software 1.0?

ML at Berkeley · 2022-09-06

This article discusses Google DeepMind's MuZero algorithm as an example of 'Software 2.0,' arguing that while deep learning surpasses traditional software, it still relies on classical computational techniques like game tree search.

0 favorites 0 likes
#deep-learning

Techniques for training large neural networks

OpenAI Blog · 2022-06-09 Cached

OpenAI presents comprehensive techniques for training large neural networks across distributed GPU clusters, covering data parallelism, pipeline parallelism, tensor parallelism, and mixture-of-experts approaches to overcome engineering and scalability challenges.

0 favorites 0 likes
#deep-learning

Introducing Triton: Open-source GPU programming for neural networks

OpenAI Blog · 2021-07-28 Cached

OpenAI releases Triton 1.0, an open-source Python-like GPU programming language that enables researchers without CUDA experience to write highly efficient GPU kernels, achieving performance on par with expert-written CUDA code in as few as 25 lines.

0 favorites 0 likes
#deep-learning

Alien Dreams: An Emerging Art Scene

ML at Berkeley · 2021-06-30 Cached

The article highlights the emerging scene of AI-generated art using OpenAI's CLIP model as a steering mechanism for generative models, showcasing various examples of text-to-image outputs.

0 favorites 0 likes
#deep-learning

Neural Module Networks for Visual Question Answering

ML at Berkeley · 2021-03-10 Cached

This article explains the Neural Module Networks (NMN) architecture from the paper 'Deep Compositional Question Answering with Neural Module Networks,' detailing how it handles the compositional structure of visual question answering tasks by decomposing questions into modular steps.

0 favorites 0 likes
← Previous
Next →
← Back to home

Submit Feedback