IntPDF
Academy
  • Guides
    • How to
    • Scholar Articles
  • Scholar Articles
    • Computer Science
      • DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
      • Personalized Wireless Federated Learning for Large Language Models
      • Consistency Guided Knowledge Retrieval and Denoising in LLMs for Zero-shot Document-level Relation Triplet Extraction
      • 3D Vision-Language Gaussian Splatting
      • Paint by Inpaint: Learning to Add Image Objects by Removing Them First
      • Raidar: geneRative AI Detection viA Rewriting
      • Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge
      • latentSplat: Autoencoding Variational Gaussians for Fast Generalizable 3D Reconstruction
      • BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks on Large Language Models
      • Do language models plan ahead for future tokens?
      • Intelligent Clinical Documentation: Harnessing Generative AI for Patient-Centric Clinical Note Generation
      • Self-Discover: Large Language Models Self-Compose Reasoning Structures
      • HyperFast: Instant Classification for Tabular Data
      • AgentReview: Exploring Peer Review Dynamics with LLM Agents
      • OpenDataLab: Empowering General Artificial Intelligence with Open Datasets
      • AgentClinic: a multimodal agent benchmark to evaluate AI in simulated clinical environments
      • Groma: Localized Visual Tokenization for Grounding Multimodal Large Language Models
      • Reinforcement Learning for Collision-free Flight Exploiting Deep Collision Encoding
      • Typos that Broke the RAG's Back: Genetic Attack on RAG Pipeline by Simulating Documents in the Wild via Low-level Perturbations
      • InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning
      • Understanding Robustness of Visual State Space Models for Image Classification
      • Ethical and social risks of harm from Language Models
      • Audio Anti-Spoofing Detection: A Survey
      • WPO: Enhancing RLHF with Weighted Preference Optimization
      • T3: Transparent Tracking Triggering for Fine-grained Overlap of Compute Collectives
      • Elephants Never Forget: Memorization and Learning of Tabular Data in Large Language Models
      • Multi-perspective Improvement of Knowledge Graph Completion with Large Language Models
      • Understanding deep learning requires rethinking generalization
      • ZipCache: Accurate and Efficient KV Cache Quantization with Salient Token Identification
      • CodeAid: Evaluating a Classroom Deployment of an LLM-based Programming Assistant that Balances Student and Educator Needs
      • RT-DETRv2: Improved Baseline with Bag-of-Freebies for Real-Time Detection Transformer
      • Is Vanilla MLP in Neural Radiance Field Enough for Few-shot View Synthesis?
      • Gemini: A Family of Highly Capable Multimodal Models
      • Large Language Model with Graph Convolution for Recommendation
      • WildGaussians: 3D Gaussian Splatting in the Wild
      • Language Models for Code Completion: A Practical Evaluation
      • Transcriptomics-guided Slide Representation Learning in Computational Pathology
      • AI and personalized learning: bridging the gap with modern educational goals
      • On the Properties of Neural Machine Translation: Encoder-Decoder Approaches
      • VCR-Graphormer: A Mini-batch Graph Transformer via Virtual Connections
      • How Reliable is Your Simulator? Analysis on the Limitations of Current LLM-based User Simulators for Conversational Recommendation
      • Meta-Prompting for Automating Zero-shot Visual Recognition with LLMs
      • Language Ranker: A Metric for Quantifying LLM Performance Across High and Low-Resource Languages
      • Visibility into AI Agents
      • Compression Represents Intelligence Linearly
      • Dual Operating Modes of In-Context Learning
      • 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations
      • OPEN TEACH: A Versatile Teleoperation System for Robotic Manipulation
      • SliM-LLM: Salience-Driven Mixed-Precision Quantization for Large Language Models
      • mDPO: Conditional Preference Optimization for Multimodal Large Language Models
      • DriVLMe: Enhancing LLM-based Autonomous Driving Agents with Embodied and Social Experiences
      • Depth-aware Test-Time Training for Zero-shot Video Object Segmentation
      • Uncertainty Quantification on Clinical Trial Outcome Prediction
      • The Interspeech 2024 Challenge on Speech Processing Using Discrete Units
      • A Survey On Text-to-3D Contents Generation In The Wild
      • MileBench: Benchmarking MLLMs in Long Context
      • Do Membership Inference Attacks Work on Large Language Models?
      • COIG-CQIA: Quality is All You Need for Chinese Instruction Fine-tuning
      • Generative Pretrained Hierarchical Transformer for Time Series Forecasting
      • How to use and interpret activation patching
      • To Generate or to Retrieve? On the Effectiveness of Artificial Contexts for Medical Open-Domain Question Answering
      • Correlation-Decoupled Knowledge Distillation for Multimodal Sentiment Analysis with Incomplete Modalities
      • Towards Interpretable Hate Speech Detection using Large Language Model-extracted Rationales
      • Transformers, parallel computation, and logarithmic depth
      • Datasheet for the Pile
      • Benchmarking Vision Language Models for Cultural Understanding
      • Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages
      • STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians
      • UniDepthV2: Universal Monocular Metric Depth Estimation Made Simpler
      • A Comprehensive Overview of Large Language Models (LLMs) for Cyber Defences: Opportunities and Directions
      • Multimodal Prompt Learning with Missing Modalities for Sentiment Analysis and Emotion Recognition
      • KAN 2.0: Kolmogorov-Arnold Networks Meet Science
      • A Survey on Hardware Accelerators for Large Language Models
      • A Comprehensive Survey on Kolmogorov Arnold Networks (KAN)
      • Flow Matching Imitation Learning for Multi-Support Manipulation
      • CRAG – Comprehensive RAG Benchmark
      • Unmasking and Quantifying Racial Bias of Large Language Models in Medical Report Generation
      • The Unreasonable Effectiveness of Eccentric Automatic Prompts
      • InFusion: Inpainting 3D Gaussians via Learning Depth Completion from Diffusion Prior
      • Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in the Era of Large Language Models
      • The PRISM Alignment Dataset: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
      • OpenTab: Advancing Large Language Models as Open-domain Table Reasoners
      • Towards Explainable, Safe Autonomous Driving with Language Embeddings for Novelty Identification and Active Learning: Framework and Experimental Analysis with Real-World Data Sets
      • Explore the Potential of CLIP for Training-Free Open Vocabulary Semantic Segmentation
      • JiuZhang3.0: Efficiently Improving Mathematical Reasoning by Training Small Data Synthesis Models
      • Probing the Creativity of Large Language Models: Can models produce divergent semantic association?
      • MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
      • LLARVA: Vision-Action Instruction Tuning Enhances Robot Learning
      • PlanAgent: A Multi-modal Large Language Agent for Closed-loop Vehicle Motion Planning
      • AutoLoRA: Automatically Tuning Matrix Ranks in Low-Rank Adaptation Based on Meta Learning
      • Spiral of Silence: How is Large Language Model Killing Information Retrieval? – A Case Study on Open Domain Question Answering
      • NuNER: Entity Recognition Encoder Pre-training via LLM-Annotated Data
      • CFPL-FAS: Class Free Prompt Learning for Generalizable Face Anti-spoofing
      • A Tale of Tails: Model Collapse as a Change of Scaling Laws
      • Prism: A Framework for Decoupling and Assessing the Capabilities of VLMs
      • Enhancing Large Language Models for Text-to-Testcase Generation
      • SemScore: Automated Evaluation of Instruction-Tuned LLMs based on Semantic Textual Similarity
      • Slow and Steady Wins the Race: Maintaining Plasticity with Hare and Tortoise Networks
      • Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
      • Diffusion Models, Image Super-Resolution And Everything: A Survey
      • Defending Large Language Models Against Jailbreak Attacks via Layer-specific Editing
      • Entropy is not Enough for Test-Time Adaptation: From the Perspective of Disentangled Factors
      • Recent Advances in Generative AI and Large Language Models: Current Status, Challenges, and Perspectives
      • Thinking Tokens for Language Modeling
      • FusionMamba: Dynamic Feature Enhancement for Multimodal Image Fusion with Mamba
      • Low-Rank Few-Shot Adaptation of Vision-Language Models
      • UniGarmentManip: A Unified Framework for Category-Level Garment Manipulation via Dense Visual Correspondence
      • Croissant: A Metadata Format for ML-Ready Datasets
      • Curriculum reinforcement learning for quantum architecture search under hardware errors
      • Applications of Deep Neural Networks with Keras
      • A Survey of Attacks on Large Vision-Language Models: Resources, Advances, and Future Trends
      • Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
      • M4GT-Bench: Evaluation Benchmark for Black-Box Machine-Generated Text Detection
      • Versatile Behavior Diffusion for Generalized Traffic Agent Simulation
      • COCONut: Modernizing COCO Segmentation
      • Exploring the Potential of Large Language Models in Self-adaptive Systems
      • Singing Voice Data Scaling-up: An Introduction to ACE-Opencpop and ACE-KiSing
      • Decentralized Multi-Robot Navigation for Autonomous Surface Vehicles with Distributional Reinforcement Learning
      • Automating Research Synthesis with Domain-Specific Large Language Model Fine-Tuning
      • Dynamic Prompt Optimizing for Text-to-Image Generation
      • What If We Recaption Billions of Web Images with LLaMA-3?
      • Open X-Embodiment: Robotic Learning Datasets and RT-X Models
      • Learning to Generate Instruction Tuning Datasets for Zero-Shot Task Adaptation
      • EIA: Environmental Injection Attack on Generalist Web Agents for Privacy Leakage
      • Top Leaderboard Ranking = Top Coding Proficiency, Always? EvoEval: Evolving Coding Benchmarks via LLM
      • TRAM: Global Trajectory and Motion of 3D Humans from in-the-wild Videos
      • A Survey on Kolmogorov-Arnold Network
      • Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?
      • DexCap: Scalable and Portable Mocap Data Collection System for Dexterous Manipulation
      • Kolmogorov-Arnold Networks are Radial Basis Function Networks
      • Spectral Networks and Locally Connected Networks on Graphs
      • DiLightNet: Fine-grained Lighting Control for Diffusion-based Image Generation
      • Personalized Language Modeling from Personalized Human Feedback
      • Research on Autonomous Robots Navigation based on Reinforcement Learning
      • Open-MAGVIT2: An Open-Source Project Toward Democratizing Auto-regressive Visual Generation
      • Dense Reward for Free in Reinforcement Learning from Human Feedback
      • Understanding Test-Time Augmentation
      • 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation
      • MuAViC: A Multilingual Audio-Visual Corpus for Robust Speech Recognition and Robust Speech-to-Text Translation
      • Multi-Object Hallucination in Vision-Language Models
      • A Novel Paradigm Boosting Translation Capabilities of Large Language Models
      • MOMENT: A Family of Open Time-series Foundation Models
      • BASS: Batched Attention-optimized Speculative Sampling
      • LLM-SR: Scientific Equation Discovery via Programming with Large Language Models
    • Economics
    • Electrical Engineering and Systems Science
      • VoCo: A Simple-yet-Effective Volume Contrastive Learning Framework for 3D Medical Image Analysis
      • Benchmarking foundation models as feature extractors for weakly-supervised computational pathology
      • ECGformer: Leveraging transformer for ECG heartbeat arrhythmia classification
    • Mathematics
      • High-fidelity single-spin shuttling in silicon
      • Exact Thermal Eigenstates of Nonintegrable Spin Chains at Infinite Temperature
    • Physics
      • Precise test of lepton flavour universality in W-boson decays into muons and electrons in pp collisions at s√=13 TeV with the ATLAS detector
      • Spin-polarized Specular Andreev Reflections in Altermagnets
      • Parametric multi-element coupling architecture for coherent and dissipative control of superconducting qubits
      • The Quantum Internet
      • Hardware-efficient quantum error correction via concatenated bosonic qubits
      • Nonreciprocal Quantum Batteries
      • High-fidelity single-spin shuttling in silicon
      • Iterative assembly of ^171Yb atom arrays with cavity-enhanced optical lattices
      • Quantum Melting of a Disordered Wigner Solid
      • Solving the strong CP problem without axions
      • DESI 2024 IV: Baryon Acoustic Oscillations from the Lyman Alpha Forest
      • Gravitational entropy is observer-dependent
      • How to factor 2048 bit RSA integers in 8 hours using 20 million noisy qubits
      • A Review of Gravitational Memory and BMS Frame Fixing in Numerical Relativity
      • The Sonora Substellar Atmosphere Models. IV. Elf Owl: Atmospheric Mixing and Chemical Disequilibrium with Varying Metallicity and C/O Ratios
      • Distinguishing oceans of water from magma on mini-Neptune K2-18b
      • Krylov complexity of density matrix operators
    • Quantitative Biology
      • Enhancing the efficiency of protein language models with minimal wet-lab data through few-shot learning
      • Out of Many, One: Designing and Scaffolding Proteins at the Scale of the Structural Universe with Genie 2
    • Quantitative Finance
    • Statistics
      • Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead
      • Linear Model and Extensions
      • Fitting Linear Mixed-Effects Models using lme4
IntPDF
  • Scholar Articles
  • Research Trends
  • Blog
Get started
Get started

Electrical Engineering and Systems Science

Benchmarking Foundation Models as Feature Extractors for Weakly-Supervised Computational Pathology →
ECGformer: Leveraging Transformer for ECG Heartbeat Arrhythmia Classification →
VoCo: A Simple-Yet-Effective Volume Contrastive Learning Framework for 3D Medical Image Analysis →
  • Privacy Policy
  • Brought to you by IntPDF