Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
72.5
TFLOPS
22
4
89
Thijs
ThijsL202
Follow
21world's profile picture
tngwwfvnqqeerpgumw's profile picture
John6666's profile picture
5 followers
·
23 following
AI & ML interests
None yet
Recent Activity
reacted
to
codelion
's
post
with 🔥
9 days ago
I recently worked on a LoRA that improves tool use in LLM. Thought the approach might interest folks here. The issue I have had when trying to use some of the local LLMs with coding agents is this: Me: "Find all API endpoints with authentication in this codebase" LLM: "You should look for @app.route decorators and check if they have auth middleware..." But I often want it to search the files and show me but the LLM doesn't trigger a tool use call. To fine-tune it for tool use I combined two data sources: 1. Magpie scenarios - 5000+ diverse tasks (bug hunting, refactoring, security audits) 2. Real execution - Ran these on actual repos (FastAPI, Django, React) to get authentic tool responses This ensures the model learns both breadth (many scenarios) and depth (real tool behavior). Tools We Taught: - `read_file` - Actually read file contents - `search_files` - Regex/pattern search across codebases - `find_definition` - Locate classes/functions - `analyze_imports` - Dependency tracking - `list_directory` - Explore structure - `run_tests` - Execute test suites Improvements: - Tool calling accuracy: 12% → 80% - Correct parameters: 8% → 87% - Multi-step tasks: 3% → 78% - End-to-end completion: 5% → 80% - Tools per task: 0.2 → 3.8 The LoRA really improves on intential tool call as an example consider the query: "Find ValueError in payment module" The response proceeds as follows: 1. Calls `search_files` with pattern "ValueError" 2. Gets 4 matches across 3 files 3. Calls `read_file` on each match 4. Analyzes context 5. Reports: "Found 3 ValueError instances: payment/processor.py:47 for invalid amount, payment/validator.py:23 for unsupported currency..." Resources: - Colab notebook https://colab.research.google.com/github/codelion/ellora/blob/main/Ellora_Recipe_3_Enhanced_Tool_Calling_and_Code_Understanding.ipynb - Model - https://huggingface.co/codelion/Llama-3.2-1B-Instruct-tool-calling-lora - GitHub - https://github.com/codelion/ellora
liked
a model
10 days ago
Lexa-B/LexaLCM_Pre3
reacted
to
dhruv3006
's
post
with 🔥
10 days ago
Pair a vision grounding model with a reasoning LLM with Cua Cua just shipped v0.4 of the Cua Agent framework with Composite Agents - you can now pair a vision/grounding model with a reasoning LLM using a simple modelA+modelB syntax. Best clicks + best plans. The problem: every GUI model speaks a different dialect. • some want pixel coordinates • others want percentages • a few spit out cursed tokens like <|loc095|> We built a universal interface that works the same across Anthropic, OpenAI, Hugging Face, etc.: agent = ComputerAgent( model="anthropic/claude-3-5-sonnet-20241022", tools=[computer] ) But here’s the fun part: you can combine models by specialization. Grounding model (sees + clicks) + Planning model (reasons + decides) → agent = ComputerAgent( model="huggingface-local/HelloKKMe/GTA1-7B+openai/gpt-4o", tools=[computer] ) This gives GUI skills to models that were never built for computer use. One handles the eyes/hands, the other the brain. Think driver + navigator working together. Two specialists beat one generalist. We’ve got a ready-to-run notebook demo - curious what combos you all will try. Github : https://github.com/trycua/cua Blog : https://www.trycua.com/blog/composite-agents
View all activity
Organizations
None yet
ThijsL202
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
New activity in
JD3GEN/JD3_Nudify_Kontext_LoRa
15 days ago
Qwen-image-edit
2
#12 opened 15 days ago by
ThijsL202
New activity in
Qwen/Qwen-Image-Edit
19 days ago
Resolution ? also as low as KONTEXT FLUX is?
👍
2
1
#12 opened 19 days ago by
AkosGmail
New activity in
zerofata/MS3.2-PaintedFantasy-24B
2 months ago
My firs impression
2
#2 opened 2 months ago by
Emalog
New activity in
Qwen/Qwen3-Embedding-0.6B-GGUF
3 months ago
Unable to load with GPU layers
👀
👍
2
1
#10 opened 3 months ago by
sambit-paul-poppulo
New activity in
h34v7/DXP-Zero-V1.2-24b-Small-Instruct
3 months ago
Why V1.2.0 instead of V1.3.0?
1
#1 opened 3 months ago by
ThijsL202
New activity in
BeaverAI/Fallen-Mistral-Small-3.1-24B-v1e-GGUF
4 months ago
curly apostrophes/feedback
#3 opened 4 months ago by
ThijsL202
New activity in
DavidAU/Qwen3-30B-A7.5B-24-Grand-Brainstorm
4 months ago
Dude's going nuts with Qwen3
5
#1 opened 4 months ago by
ponzles
New activity in
DavidAU/Llama3.1-MOE-4X8B-Gated-IQ-Multi-Tier-COGITO-Deep-Reasoning-32B-GGUF
5 months ago
Shorter reasoning?
2
#2 opened 5 months ago by
ThijsL202
New activity in
lucyknada/Gryphe_Pantheon-RP-1.8-24b-Small-3.1-exl2
5 months ago
Gibberish output using tabbyapi
7
#1 opened 5 months ago by
ThijsL202
New activity in
ArliAI/QwQ-32B-ArliAI-RpR-v1
5 months ago
Is the moral limit of this model high?
3
#1 opened 5 months ago by
Ankahia
New activity in
LatitudeGames/Wayfarer-12B
7 months ago
Bigger models of this?
👀
👍
2
3
#6 opened 7 months ago by
Adzeiros
New activity in
TheDrummer/Cydonia-24B-v2
7 months ago
recommended sampler settings
👀
1
2
#4 opened 7 months ago by
plowthat1998
New activity in
DavidAU/AI_Autocorrect__Auto-Creative-Enhancement__Auto-Low-Quant-Optimization__gguf-exl2-hqq-SOFTWARE
7 months ago
Sillytavern not Initilazing
5
#3 opened 7 months ago by
Skydea
New activity in
DavidAU/MN-Dark-Planet-Kaboom-21B-GGUF
12 months ago
Garbage output? (Q6_k)
7
#1 opened 12 months ago by
ThijsL202
New activity in
TheDrummer/Tiger-Gemma-9B-v1
about 1 year ago
Differences between Tiger Gemma, Smegmma and Broken Gemma
22
#1 opened about 1 year ago by
isr431
New activity in
BeaverLegacy/Smegmma-Deluxe-9B-v1
about 1 year ago
My quantizations.
🤗
2
5
#1 opened about 1 year ago by
ZeroWw
My quantizations.
🤗
2
5
#1 opened about 1 year ago by
ZeroWw
Load more