Computer Science, Computers and Society
Author: LLama 2 7B Chat
Page 66/179
LLaMA-2, the next generation of LLaMA. Meta trained and released LLaMA-2 in three model sizes: 7, 13, and 70 billion parameters. The model architecture remains largely unchanged from that of LLaMA-1 models, but 40% more data was used to train the foundational models. The accompanying preprint also mentions a model with 34B parameters that might be released in the future upon satisfying safety targets.
Computation and Language, Computer Science
Re-Evaluating Factual Consistency Evaluation in Natural Language Processing
Electrical Engineering and Systems Science, Systems and Control
Machine Learning for Signal Detection: A Unified Approach
Computer Science, Computer Vision and Pattern Recognition
Mitigating Distribution Shifts in Active Perception with Simulation
Training Policies for Gentle Manipulation in Robot-Assisted Surgery via Imitation Learning
Electrical Engineering and Systems Science, Systems and Control
Robust Chaos-Based Missile Guidance Law with Adaptive Parameters
Comparing Weights of Attention Mechanisms in Multi-Label Classification
Computer Science, Information Theory
Designing and Interpreting Probes with Control Tasks: A Comprehensive Review
Computation and Language, Computer Science
Improving LLM Predictions via Demonstrations and Sampling
Computable Dimension, Randomness, and Normality: A Study of Effective Representations
Computer Science, Machine Learning
Improved Deep Learning Models for Efficient Face Anti-Spoofing
Computer Science, Computer Science and Game Theory
The Value of Difficulty in Social Games
Computer Science, Networking and Internet Architecture
Exponential Growth in Network Load Factors: A Computationally Efficient Solution
Computer Science, Computer Vision and Pattern Recognition
Assessing Surgical Skills through Tree-Based Gaussian Process Classifier
Computer Science, Machine Learning
Implicit Opinion in AI: Understanding Alice’s Perspective
Computer Science, Human-Computer Interaction
Enhancing Modeling Capacity with Stacked Pairwise Attention Layers: A Comparative Study of SWAN and Transformer
Computer Science, Computer Vision and Pattern Recognition
Improving Region-Level Captioning with Osprey-7B: A Quantitative Comparison Study.
Computation and Language, Computer Science
Enhanced Gloss2Text Model for Adaptive Translation of Visual-Grounded Text
Computer Science, Machine Learning
Detecting Concept Drift in Dependent Data using Dynamic Adaptive Window Independence Drift Detection
Computer Science, Computer Vision and Pattern Recognition