stereoplegic 's Collections Embeddings
updated
Towards General Text Embeddings with Multi-stage Contrastive Learning
Paper
• 2308.03281
• Published
• 3
NEFTune: Noisy Embeddings Improve Instruction Finetuning
Paper
• 2310.05914
• Published
• 14
EELBERT: Tiny Models through Dynamic Embeddings
Paper
• 2310.20144
• Published
• 3
Dynamic Word Embeddings for Evolving Semantic Discovery
Paper
• 1703.00607
• Published
• 1
Paper
• 1702.08359
• Published
• 1
Unleashing the Power of Pre-trained Language Models for Offline
Reinforcement Learning
Paper
• 2310.20587
• Published
• 18
Understanding and Improving Information Transfer in Multi-Task Learning
Paper
• 2005.00944
• Published
• 1
Split, Encode and Aggregate for Long Code Search
Paper
• 2208.11271
• Published
• 1
TEAL: Tokenize and Embed ALL for Multi-modal Large Language Models
Paper
• 2311.04589
• Published
• 21
Frustratingly Simple Memory Efficiency for Pre-trained Language Models
via Dynamic Embedding Pruning
Paper
• 2309.08708
• Published
• 3
CokeBERT: Contextual Knowledge Selection and Embedding towards Enhanced
Pre-Trained Language Models
Paper
• 2009.13964
• Published
• 1
Plug-and-Play Knowledge Injection for Pre-trained Language Models
Paper
• 2305.17691
• Published
• 1
Plug-and-Play Document Modules for Pre-trained Models
Paper
• 2305.17660
• Published
• 1
Identifying Linear Relational Concepts in Large Language Models
Paper
• 2311.08968
• Published
• 1
Can the Inference Logic of Large Language Models be Disentangled into
Symbolic Concepts?
Paper
• 2304.01083
• Published
• 1
Sub-Sentence Encoder: Contrastive Learning of Propositional Semantic
Representations
Paper
• 2311.04335
• Published
• 1
Retrieve Anything To Augment Large Language Models
Paper
• 2310.07554
• Published
• 6
Nomic Embed: Training a Reproducible Long Context Text Embedder
Paper
• 2402.01613
• Published
• 15
Improving Text Embeddings with Large Language Models
Paper
• 2401.00368
• Published
• 82
Uncovering hidden geometry in Transformers via disentangling position
and context
Paper
• 2310.04861
• Published
Char2Subword: Extending the Subword Embedding Space Using Robust
Character Compositionality
Paper
• 2010.12730
• Published
CharBERT: Character-aware Pre-trained Language Model
Paper
• 2011.01513
• Published
Training Multilingual Pre-trained Language Model with Byte-level
Subwords
Paper
• 2101.09469
• Published
Neural Machine Translation with Byte-Level Subwords
Paper
• 1909.03341
• Published
• 1
byteSteady: Fast Classification Using Byte-Level n-Gram Embeddings
Paper
• 2106.13302
• Published
MAGNET: Improving the Multilingual Fairness of Language Models with
Adaptive Gradient-Based Tokenization
Paper
• 2407.08818
• Published
OFA: A Framework of Initializing Unseen Subword Embeddings for Efficient
Large-scale Multilingual Continued Pretraining
Paper
• 2311.08849
• Published
• 6