Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
45.5
TFLOPS
160
13
62
Jean Louis
JLouisBiz
Follow
Kseniase's profile picture
chaksaray's profile picture
as-cle-bert's profile picture
76 followers
·
114 following
https://www.StartYourOwnGoldMine.com
YourOwnGoldMine
gnusupport
AI & ML interests
- LLM for sales, marketing, promotion - LLM for Website Revision System - increasing quality of communication with customers - helping clients access information faster - saving people from financial troubles
Recent Activity
reacted
to
mitkox
's
post
with 👍
about 15 hours ago
Got to 1199.8 tokens/sec with Devstral Small -2 on my desktop GPU workstation. vLLM nightly. Works out of the box with Mistral Vibe. Next is time to test the big one.
replied
to
mitkox
's
post
about 15 hours ago
Got to 1199.8 tokens/sec with Devstral Small -2 on my desktop GPU workstation. vLLM nightly. Works out of the box with Mistral Vibe. Next is time to test the big one.
replied
to
melvindave
's
post
about 16 hours ago
Currently having a blast learning the transformers library. I noticed that model cards usually have Transformers code as usage examples. So I tried to figure out how to load a model just using the transformers library without using ollama, lmstudio, or llamacpp. Learned how to install dependencies required to make it work like pytorch and CUDA. I also used Conda for python environment dependencies. Once I got the model loaded and sample inference working, I made an API to serve it. I know it's very basic stuff for machine learning experts here in HF but I'm completely new to this so I'm happy to get it working! Model used: https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct GPU: NVIDIA GeForce RTX 3090 Here's the result of my experimentation
View all activity
Organizations
JLouisBiz
's models
2
Sort: Recently updated
JLouisBiz/Qwen3-0.6B-GGUF-for-24GB-VRAM
Text Generation
•
0.8B
•
Updated
Apr 29
•
66
•
2
JLouisBiz/Qwen3-1.7B-for-24GB-VRAM
Text Generation
•
2B
•
Updated
Apr 29
•
9