clip-tagger Model

This is a personalized audio tagging model based on CLAP (Contrastive Language-Audio Pre-training). It extends the base Xenova/clap-htsat-unfused model with user feedback and custom tags.

Model Description

  • Base Model: Xenova/clap-htsat-unfused
  • Framework: Transformers.js compatible
  • Training: User feedback and custom tag integration
  • Use Case: Personalized audio content tagging

Usage

import { CLAPProcessor } from './clapProcessor.js';
import { LocalClassifier } from './localClassifier.js';

// Load the model
const processor = new CLAPProcessor();
const classifier = new LocalClassifier();
classifier.loadModel(); // Loads from localStorage or model files

// Process audio
const tags = await processor.processAudio(audioBuffer);
const personalizedTags = classifier.predictAll(features, candidateTags);

Files

  • localClassifier.js - Local classifier implementation
  • clapProcessor.js - CLAP model wrapper
  • userFeedbackStore.js - User feedback storage system
  • model-config.json - Model configuration
  • example-usage.html - Usage example

Links

Training Data

This model learns from user corrections and custom tags. The base CLAP model provides initial audio understanding, while the local classifier adapts to user preferences.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for sohei1l/clip-tagger

Finetuned
(1)
this model