Computation and Language, Computer Science
Author: LLama 2 7B Chat
Page 121/179
LLaMA-2, the next generation of LLaMA. Meta trained and released LLaMA-2 in three model sizes: 7, 13, and 70 billion parameters. The model architecture remains largely unchanged from that of LLaMA-1 models, but 40% more data was used to train the foundational models. The accompanying preprint also mentions a model with 34B parameters that might be released in the future upon satisfying safety targets.
Electrical Engineering and Systems Science, Image and Video Processing
Mitigating Bias in Synthetic Data Generation for Surgical Simulation
Computer Science, Machine Learning
Incremental Improvements in Neural Networks: A Gentle Introduction to Graph Neural Networks
Computation and Language, Computer Science
Uncovering Physician Fatigue Through Clinical Notes
Computer Science, Human-Computer Interaction
Enhancing Event Understanding with Provenance and Filtering Tools
Computer Science, Machine Learning
Uncovering Incidental Polysemanticity in Deep Neural Networks
Computer Science, Machine Learning
Evolutionary Machine Learning: A Comprehensive Survey
Artificial Intelligence, Computer Science
Evaluating Agents using Social Choice Theory: A Comparative Analysis of Prompts
Computer Science, Machine Learning
Logically Consistent Reasoning with Recursive Expla-nations
Computer Science, Computers and Society
Improving Reading Comprehension Question Generation with Distractors
Computer Science, Machine Learning
Essential Characteristics of Tasks for Robustness in Transfer Learning
Computer Science, Machine Learning
Preventing Highway Congestion with Autonomous Vehicles and Reinforcement Learning
Computer Science, Computer Vision and Pattern Recognition
Hierarchical Text-Conditional Image Generation with CLIP Latents
Artificial Intelligence, Computer Science
Imitation Learning: The Path to Humanoid Robots?
Computation and Language, Computer Science
The Hidden Truth Behind Large Language Models in Scientific Publishing
Numerical Analysis of Differential Conditional Dependency Structures in Synthetic and Real-World Networks
Computer Science, Machine Learning
Underspecification in Deep Reinforcement Learning: A Study on Goal Misgeneralization
Computer Science, Computer Vision and Pattern Recognition
Efficient Geometry-Aware 3D Generation with Diffusion Models
Computer Science, Machine Learning
Similarity-Based Knowledge Transfer for Cross-Domain Reinforcement Learning
Computation and Language, Computer Science