Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
57
18
134
SeanLee
SeanLee97
Follow
Xurinth's profile picture
ChengzudongLingnan's profile picture
nileshhanotia's profile picture
54 followers
Β·
28 following
seanlee97
AI & ML interests
NLP; Information Retrieval; Information Extraction
Recent Activity
liked
a model
about 4 hours ago
tencent/HY-Embodied-0.5
reacted
to
their
post
with π
about 6 hours ago
Our lab recently released a paper where we introduce ShadowPEFT, a new Parameter-Efficient Fine-Tuning (PEFT) paradigm tailored for edge computing scenarios. Unlike traditional approaches such as LoRA and its variants, which inject trainable parameters directly into the weights of Transformer, requiring tight coupling with the backbone. ShadowPEFT instead enhances the frozen large base model by adding a lightweight, centralized, pretrainable, and detachable Shadow network. This shadow network operates in parallel with the base model, delivering learned corrections to each decoder layer. Because the shadow module is architecturally decoupled from the backbone, it can be independently trained, stored, and deployed, benefiting edge computing scenarios and edge-cloud collaboration computing. - HF Paper: https://huggingface.co/papers/2604.19254 - GitHub: https://github.com/ShadowLLM/shadow-peft - HF Collection: https://huggingface.co/collections/shadow-llm/shadow-peft-models
reacted
to
their
post
with π
about 6 hours ago
Our lab recently released a paper where we introduce ShadowPEFT, a new Parameter-Efficient Fine-Tuning (PEFT) paradigm tailored for edge computing scenarios. Unlike traditional approaches such as LoRA and its variants, which inject trainable parameters directly into the weights of Transformer, requiring tight coupling with the backbone. ShadowPEFT instead enhances the frozen large base model by adding a lightweight, centralized, pretrainable, and detachable Shadow network. This shadow network operates in parallel with the base model, delivering learned corrections to each decoder layer. Because the shadow module is architecturally decoupled from the backbone, it can be independently trained, stored, and deployed, benefiting edge computing scenarios and edge-cloud collaboration computing. - HF Paper: https://huggingface.co/papers/2604.19254 - GitHub: https://github.com/ShadowLLM/shadow-peft - HF Collection: https://huggingface.co/collections/shadow-llm/shadow-peft-models
View all activity
Organizations
SeanLee97
's datasets
4
Sort:Β Recently updated
SeanLee97/nli_format_a_for_angle_test
Viewer
β’
Updated
Sep 29, 2024
β’
100
β’
8
SeanLee97/all_nli_angle_format_a
Viewer
β’
Updated
Jun 28, 2024
β’
962k
β’
415
SeanLee97/nli_for_simcse
Viewer
β’
Updated
Apr 19, 2024
β’
276k
β’
1.86k
SeanLee97/all_nli_angle_format_b
Viewer
β’
Updated
Apr 8, 2024
β’
481k
β’
10
β’
1