Mathematics, Numerical Analysis
Author: LLama 2 7B Chat
Page 34/179
LLaMA-2, the next generation of LLaMA. Meta trained and released LLaMA-2 in three model sizes: 7, 13, and 70 billion parameters. The model architecture remains largely unchanged from that of LLaMA-1 models, but 40% more data was used to train the foundational models. The accompanying preprint also mentions a model with 34B parameters that might be released in the future upon satisfying safety targets.
Computation and Language, Computer Science
Efficiently Processing Complex Biomedical Text with Domain-Aware Models
Quantitative Finance, Risk Management
Orthogonal Network Interventions for Acceptability
Computer Science, Software Engineering
Shipwright: A Human-in-the-Loop System for Dockerfile Repair
novel block-wise Metropolization
Recursive Summarization Enables Long-Term Dialogue Memory in Large Language Models
Computer Science, Machine Learning
Uncovering Anomalies with Copula-Based Outlier Detection
Computer Science, Software Engineering
Measuring Software Energy Consumption: A Comprehensive Analysis
Fine-Tuning Large Language Models for Domain-Specific Tasks: A Comparative Study
A Comprehensive Review of Deep Learning for Object Segmentation in Robotics
Scaling Rewards for Efficient Reinforcement Learning
Artificial Intelligence, Computer Science
Robust Reinforcement Learning: A New Frontier in MDPs
Computer Science, Computer Vision and Pattern Recognition
Unified Text-to-Text Transformer
Computer Science, Information Theory
Semantics-Aware Information Dissemination for Efficient Network Energy Management
Computer Science, Information Theory
Correcting Codes in Polynomial Rings: A Review of Quantum Error Correction Techniques
Computer Science, Machine Learning
Fed-CO2: A Communication-Efficient Federated Learning Algorithm for Heterogeneous Data
Computer Science, Human-Computer Interaction
Improving End-to-End Task-Oriented Dialog Systems with Asynchronous Coordination
Computer Science, Machine Learning
Efficient Label Acquisition for Machine Learning Prediction Tasks
Computer Science, Networking and Internet Architecture
Task-Oriented Communications: Efficient Data Delivery for Machine Learning Tasks
Computation and Language, Computer Science