id
stringlengths
9
104
author
stringlengths
3
36
task_category
stringclasses
32 values
tags
sequencelengths
1
4.05k
created_time
unknowndate
2022-03-02 23:29:04
2025-03-18 02:34:30
last_modified
stringdate
2021-02-13 00:06:56
2025-03-18 09:30:19
downloads
int64
0
15.6M
likes
int64
0
4.86k
README
stringlengths
44
1.01M
matched_bigbio_names
sequencelengths
1
8
ostapeno/indepexp_adauniNeo1B_sciq_Multiple_Choice_sub05_3ep
ostapeno
null
[ "region:us" ]
"2024-01-12T15:25:03Z"
2024-01-12T16:23:42+00:00
0
0
--- {} --- Number of experts present in the library: 3 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | sciq_Multiple_Choice | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice | lora | | sciq_Multiple_Choice_v1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice | lora | | sciq_Multiple_Choice_v2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice | lora | Last updated on: 2024-01-12 16:23:40+00:00
[ "SCIQ" ]
ntc-ai/SDXL-LoRA-slider.the-starry-night
ntc-ai
text-to-image
[ "diffusers", "text-to-image", "stable-diffusion-xl", "lora", "template:sd-lora", "template:sdxl-lora", "sdxl-sliders", "ntcai.xyz-sliders", "concept", "en", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:mit", "region:us" ]
"2024-01-13T00:21:26Z"
2024-01-13T00:21:29+00:00
0
0
--- base_model: stabilityai/stable-diffusion-xl-base-1.0 language: - en license: mit tags: - text-to-image - stable-diffusion-xl - lora - template:sd-lora - template:sdxl-lora - sdxl-sliders - ntcai.xyz-sliders - concept - diffusers thumbnail: images/evaluate/the starry night.../the starry night_17_3.0.png widget: - text: the starry night output: url: images/the starry night_17_3.0.png - text: the starry night output: url: images/the starry night_19_3.0.png - text: the starry night output: url: images/the starry night_20_3.0.png - text: the starry night output: url: images/the starry night_21_3.0.png - text: the starry night output: url: images/the starry night_22_3.0.png inference: false instance_prompt: the starry night --- # ntcai.xyz slider - the starry night (SDXL LoRA) | Strength: -3 | Strength: 0 | Strength: 3 | | --- | --- | --- | | <img src="images/the starry night_17_-3.0.png" width=256 height=256 /> | <img src="images/the starry night_17_0.0.png" width=256 height=256 /> | <img src="images/the starry night_17_3.0.png" width=256 height=256 /> | | <img src="images/the starry night_19_-3.0.png" width=256 height=256 /> | <img src="images/the starry night_19_0.0.png" width=256 height=256 /> | <img src="images/the starry night_19_3.0.png" width=256 height=256 /> | | <img src="images/the starry night_20_-3.0.png" width=256 height=256 /> | <img src="images/the starry night_20_0.0.png" width=256 height=256 /> | <img src="images/the starry night_20_3.0.png" width=256 height=256 /> | ## Download Weights for this model are available in Safetensors format. ## Trigger words You can apply this LoRA with trigger words for additional effect: ``` the starry night ``` ## Use in diffusers ```python from diffusers import StableDiffusionXLPipeline from diffusers import EulerAncestralDiscreteScheduler import torch pipe = StableDiffusionXLPipeline.from_single_file("https://huggingface.co/martyn/sdxl-turbo-mario-merge-top-rated/blob/main/topRatedTurboxlLCM_v10.safetensors") pipe.to("cuda") pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) # Load the LoRA pipe.load_lora_weights('ntc-ai/SDXL-LoRA-slider.the-starry-night', weight_name='the starry night.safetensors', adapter_name="the starry night") # Activate the LoRA pipe.set_adapters(["the starry night"], adapter_weights=[2.0]) prompt = "medieval rich kingpin sitting in a tavern, the starry night" negative_prompt = "nsfw" width = 512 height = 512 num_inference_steps = 10 guidance_scale = 2 image = pipe(prompt, negative_prompt=negative_prompt, width=width, height=height, guidance_scale=guidance_scale, num_inference_steps=num_inference_steps).images[0] image.save('result.png') ``` ## Support the Patreon If you like this model please consider [joining our Patreon](https://www.patreon.com/NTCAI). By joining our Patreon, you'll gain access to an ever-growing library of over 1070+ unique and diverse LoRAs, covering a wide range of styles and genres. You'll also receive early access to new models and updates, exclusive behind-the-scenes content, and the powerful LoRA slider creator, allowing you to craft your own custom LoRAs and experiment with endless possibilities. Your support on Patreon will allow us to continue developing and refining new models. ## Other resources - [CivitAI](https://civitai.com/user/ntc) - Follow ntc on Civit for even more LoRAs - [ntcai.xyz](https://ntcai.xyz) - See ntcai.xyz to find more articles and LoRAs
[ "CRAFT" ]
rpayanm/stable-ts
rpayanm
null
[ "en", "license:mit", "region:us" ]
"2024-01-13T15:46:55Z"
2024-01-13T16:55:55+00:00
0
0
--- language: - en license: mit --- # Stabilizing Timestamps for Whisper This library modifies [Whisper](https://github.com/openai/whisper) to produce more reliable timestamps and extends its functionality. https://github.com/jianfch/stable-ts/assets/28970749/7adf0540-3620-4b2b-b2d4-e316906d6dfa * [Setup](#setup) * [Usage](#usage) * [Transcribe](#transcribe) * [Output](#output) * [Alignment](#alignment) * [Adjustments](#adjustments) * [Refinement](#refinement) * [Regrouping Words](#regrouping-words) * [Editing](#editing) * [Locating Words](#locating-words) * [Silence Suppression](#silence-suppression) * [Tips](#tips) * [Visualizing Suppression](#visualizing-suppression) * [Encode Comparison](#encode-comparison) * [Use with any ASR](#any-asr) * [Quick 1.X → 2.X Guide](#quick-1x--2x-guide) ## Setup ``` pip install -U stable-ts ``` To install the latest commit: ``` pip install -U git+https://github.com/jianfch/stable-ts.git ``` ## Usage ### Transcribe ```python import stable_whisper model = stable_whisper.load_model('base') result = model.transcribe('audio.mp3') result.to_srt_vtt('audio.srt') ``` <details> <summary>CLI</summary> ```commandline stable-ts audio.mp3 -o audio.srt ``` </details> Docstrings: <details> <summary>load_model()</summary> Load an instance if :class:`whisper.model.Whisper`. Parameters ---------- name : {'tiny', 'tiny.en', 'base', 'base.en', 'small', 'small.en', 'medium', 'medium.en', 'large-v1', 'large-v2', 'large-v3', or 'large'} One of the official model names listed by :func:`whisper.available_models`, or path to a model checkpoint containing the model dimensions and the model state_dict. device : str or torch.device, optional PyTorch device to put the model into. download_root : str, optional Path to download the model files; by default, it uses "~/.cache/whisper". in_memory : bool, default False Whether to preload the model weights into host memory. cpu_preload : bool, default True Load model into CPU memory first then move model to specified device to reduce GPU memory usage when loading model dq : bool, default False Whether to apply Dynamic Quantization to model to reduced memory usage and increase inference speed but at the cost of a slight decrease in accuracy. Only for CPU. Returns ------- model : "Whisper" The Whisper ASR model instance. Notes ----- The overhead from ``dq = True`` might make inference slower for models smaller than 'large'. </details> <details> <summary>transcribe()</summary> Transcribe audio using Whisper. This is a modified version of :func:`whisper.transcribe.transcribe` with slightly different decoding logic while allowing additional preprocessing and postprocessing. The preprocessing performed on the audio includes: isolating voice / removing noise with Demucs and low/high-pass filter. The postprocessing performed on the transcription result includes: adjusting timestamps with VAD and custom regrouping segments based punctuation and speech gaps. Parameters ---------- model : whisper.model.Whisper An instance of Whisper ASR model. audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is :class:`numpy.ndarray` or :class:`torch.Tensor`, the audio must be already at sampled to 16kHz. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. temperature : float or iterable of float, default (0.0, 0.2, 0.4, 0.6, 0.8, 1.0) Temperature for sampling. It can be a tuple of temperatures, which will be successfully used upon failures according to either ``compression_ratio_threshold`` or ``logprob_threshold``. compression_ratio_threshold : float, default 2.4 If the gzip compression ratio is above this value, treat as failed. logprob_threshold : float, default -1 If the average log probability over sampled tokens is below this value, treat as failed no_speech_threshold : float, default 0.6 If the no_speech probability is higher than this value AND the average log probability over sampled tokens is below ``logprob_threshold``, consider the segment as silent condition_on_previous_text : bool, default True If ``True``, the previous output of the model is provided as a prompt for the next window; disabling may make the text inconsistent across windows, but the model becomes less prone to getting stuck in a failure loop, such as repetition looping or timestamps going out of sync. initial_prompt : str, optional Text to provide as a prompt for the first window. This can be used to provide, or "prompt-engineer" a context for transcription, e.g. custom vocabularies or proper nouns to make it more likely to predict those word correctly. word_timestamps : bool, default True Extract word-level timestamps using the cross-attention pattern and dynamic time warping, and include the timestamps for each word in each segment. Disabling this will prevent segments from splitting/merging properly. regroup : bool or str, default True, meaning the default regroup algorithm String for customizing the regrouping algorithm. False disables regrouping. Ignored if ``word_timestamps = False``. ts_num : int, default 0, meaning disable this option Number of extra timestamp inferences to perform then use average of these extra timestamps. An experimental option that might hurt performance. ts_noise : float, default 0.1 Percentage of noise to add to audio_features to perform inferences for ``ts_num``. suppress_silence : bool, default True Whether to enable timestamps adjustments based on the detected silence. suppress_word_ts : bool, default True Whether to adjust word timestamps based on the detected silence. Only enabled if ``suppress_silence = True``. use_word_position : bool, default True Whether to use position of the word in its segment to determine whether to keep end or start timestamps if adjustments are required. If it is the first word, keep end. Else if it is the last word, keep the start. q_levels : int, default 20 Quantization levels for generating timestamp suppression mask; ignored if ``vad = true``. Acts as a threshold to marking sound as silent. Fewer levels will increase the threshold of volume at which to mark a sound as silent. k_size : int, default 5 Kernel size for avg-pooling waveform to generate timestamp suppression mask; ignored if ``vad = true``. Recommend 5 or 3; higher sizes will reduce detection of silence. time_scale : float, optional Factor for scaling audio duration for inference. Greater than 1.0 'slows down' the audio, and less than 1.0 'speeds up' the audio. None is same as 1.0. A factor of 1.5 will stretch 10s audio to 15s for inference. This increases the effective resolution of the model but can increase word error rate. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo. https://github.com/facebookresearch/demucs. demucs_output : str, optional Path to save the vocals isolated by Demucs as WAV file. Ignored if ``demucs = False``. Demucs must be installed to use. Official repo. https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. vad : bool, default False Whether to use Silero VAD to generate timestamp suppression mask. Silero VAD requires PyTorch 1.12.0+. Official repo, https://github.com/snakers4/silero-vad. vad_threshold : float, default 0.35 Threshold for detecting speech with Silero VAD. Low threshold reduces false positives for silence detection. vad_onnx : bool, default False Whether to use ONNX for Silero VAD. min_word_dur : float, default 0.1 Shortest duration each word is allowed to reach for silence suppression. nonspeech_error : float, default 0.3 Relative error of non-speech sections that appear in between a word for silence suppression. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. prepend_punctuations : str, default '"\'“¿([{-)' Punctuations to prepend to next word. append_punctuations : str, default '.。,,!!??::”)]}、)' Punctuations to append to previous word. mel_first : bool, default False Process entire audio track into log-Mel spectrogram first instead in chunks. Used if odd behavior seen in stable-ts but not in whisper, but use significantly more memory for long audio. split_callback : Callable, optional Custom callback for grouping tokens up with their corresponding words. The callback must take two arguments, list of tokens and tokenizer. The callback returns a tuple with a list of words and a corresponding nested list of tokens. suppress_ts_tokens : bool, default False Whether to suppress timestamp tokens during inference for timestamps are detected at silent. Reduces hallucinations in some cases, but also prone to ignore disfluencies and repetitions. This option is ignored if ``suppress_silence = False``. gap_padding : str, default ' ...' Padding prepend to each segments for word timing alignment. Used to reduce the probability of model predicting timestamps earlier than the first utterance. only_ffmpeg : bool, default False Whether to use only FFmpeg (instead of not yt-dlp) for URls max_instant_words : float, default 0.5 If percentage of instantaneous words in a segment exceed this amount, the segment is removed. avg_prob_threshold: float or None, default None Transcribe the gap after the previous word and if the average word proababiliy of a segment falls below this value, discard the segment. If ``None``, skip transcribing the gap to reduce chance of timestamps starting before the next utterance. progress_callback : Callable, optional A function that will be called when transcription progress is updated. The callback need two parameters. The first parameter is a float for seconds of the audio that has been transcribed. The second parameter is a float for total duration of audio in seconds. ignore_compatibility : bool, default False Whether to ignore warnings for compatibility issues with the detected Whisper version. decode_options Keyword arguments to construct class:`whisper.decode.DecodingOptions` instances. Returns ------- stable_whisper.result.WhisperResult All timestamps, words, probabilities, and other data from the transcription of ``audio``. See Also -------- stable_whisper.non_whisper.transcribe_any : Return :class:`stable_whisper.result.WhisperResult` containing all the data from transcribing audio with unmodified :func:`whisper.transcribe.transcribe` with preprocessing and postprocessing. stable_whisper.whisper_word_level.load_faster_whisper.faster_transcribe : Return :class:`stable_whisper.result.WhisperResult` containing all the data from transcribing audio with :meth:`faster_whisper.WhisperModel.transcribe` with preprocessing and postprocessing. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3', vad=True) >>> result.to_srt_vtt('audio.srt') Saved: audio.srt </details> <details> <summary>transcribe_minimal()</summary> Transcribe audio using Whisper. This is uses the original whisper transcribe function, :func:`whisper.transcribe.transcribe`, while still allowing additional preprocessing and postprocessing. The preprocessing performed on the audio includes: isolating voice / removing noise with Demucs and low/high-pass filter. The postprocessing performed on the transcription result includes: adjusting timestamps with VAD and custom regrouping segments based punctuation and speech gaps. Parameters ---------- model : whisper.model.Whisper An instance of Whisper ASR model. audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is ``numpy.ndarray`` or ``torch.Tensor``, the audio must be already at sampled to 16kHz. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. word_timestamps : bool, default True Extract word-level timestamps using the cross-attention pattern and dynamic time warping, and include the timestamps for each word in each segment. Disabling this will prevent segments from splitting/merging properly. regroup : bool or str, default True, meaning the default regroup algorithm String for customizing the regrouping algorithm. False disables regrouping. Ignored if ``word_timestamps = False``. suppress_silence : bool, default True Whether to enable timestamps adjustments based on the detected silence. suppress_word_ts : bool, default True Whether to adjust word timestamps based on the detected silence. Only enabled if ``suppress_silence = True``. use_word_position : bool, default True Whether to use position of the word in its segment to determine whether to keep end or start timestamps if adjustments are required. If it is the first word, keep end. Else if it is the last word, keep the start. q_levels : int, default 20 Quantization levels for generating timestamp suppression mask; ignored if ``vad = true``. Acts as a threshold to marking sound as silent. Fewer levels will increase the threshold of volume at which to mark a sound as silent. k_size : int, default 5 Kernel size for avg-pooling waveform to generate timestamp suppression mask; ignored if ``vad = true``. Recommend 5 or 3; higher sizes will reduce detection of silence. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_output : str, optional Path to save the vocals isolated by Demucs as WAV file. Ignored if ``demucs = False``. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. vad : bool, default False Whether to use Silero VAD to generate timestamp suppression mask. Silero VAD requires PyTorch 1.12.0+. Official repo, https://github.com/snakers4/silero-vad. vad_threshold : float, default 0.35 Threshold for detecting speech with Silero VAD. Low threshold reduces false positives for silence detection. vad_onnx : bool, default False Whether to use ONNX for Silero VAD. min_word_dur : float, default 0.1 Shortest duration each word is allowed to reach for silence suppression. nonspeech_error : float, default 0.3 Relative error of non-speech sections that appear in between a word for silence suppression. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. only_ffmpeg : bool, default False Whether to use only FFmpeg (instead of not yt-dlp) for URls options Additional options used for :func:`whisper.transcribe.transcribe` and :func:`stable_whisper.non_whisper.transcribe_any`. Returns ------- stable_whisper.result.WhisperResult All timestamps, words, probabilities, and other data from the transcription of ``audio``. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe_minimal('audio.mp3', vad=True) >>> result.to_srt_vtt('audio.srt') Saved: audio.srt </details> <br> <details> <summary>faster-whisper</summary> Use with [faster-whisper](https://github.com/guillaumekln/faster-whisper): ```python model = stable_whisper.load_faster_whisper('base') result = model.transcribe_stable('audio.mp3') ``` ```commandline stable-ts audio.mp3 -o audio.srt -fw ``` Docstring: <details> <summary>load_faster_whisper()</summary> Load an instance of :class:`faster_whisper.WhisperModel`. Parameters ---------- model_size_or_path : {'tiny', 'tiny.en', 'base', 'base.en', 'small', 'small.en', 'medium', 'medium.en', 'large-v1', 'large-v2', 'large-v3', or 'large'} Size of the model. model_init_options Additional options to use for initialization of :class:`faster_whisper.WhisperModel`. Returns ------- faster_whisper.WhisperModel A modified instance with :func:`stable_whisper.whisper_word_level.load_faster_whisper.faster_transcribe` assigned to :meth:`faster_whisper.WhisperModel.transcribe_stable`. </details> <details> <summary>transcribe_stable()</summary> Transcribe audio using faster-whisper (https://github.com/guillaumekln/faster-whisper). This is uses the transcribe method from faster-whisper, :meth:`faster_whisper.WhisperModel.transcribe`, while still allowing additional preprocessing and postprocessing. The preprocessing performed on the audio includes: isolating voice / removing noise with Demucs and low/high-pass filter. The postprocessing performed on the transcription result includes: adjusting timestamps with VAD and custom regrouping segments based punctuation and speech gaps. Parameters ---------- model : faster_whisper.WhisperModel The faster-whisper ASR model instance. audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is :class:`numpy.ndarray` or :class:`torch.Tensor`, the audio must be already at sampled to 16kHz. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. word_timestamps : bool, default True Extract word-level timestamps using the cross-attention pattern and dynamic time warping, and include the timestamps for each word in each segment. Disabling this will prevent segments from splitting/merging properly. regroup : bool or str, default True, meaning the default regroup algorithm String for customizing the regrouping algorithm. False disables regrouping. Ignored if ``word_timestamps = False``. suppress_silence : bool, default True Whether to enable timestamps adjustments based on the detected silence. suppress_word_ts : bool, default True Whether to adjust word timestamps based on the detected silence. Only enabled if ``suppress_silence = True``. use_word_position : bool, default True Whether to use position of the word in its segment to determine whether to keep end or start timestamps if adjustments are required. If it is the first word, keep end. Else if it is the last word, keep the start. q_levels : int, default 20 Quantization levels for generating timestamp suppression mask; ignored if ``vad = true``. Acts as a threshold to marking sound as silent. Fewer levels will increase the threshold of volume at which to mark a sound as silent. k_size : int, default 5 Kernel size for avg-pooling waveform to generate timestamp suppression mask; ignored if ``vad = true``. Recommend 5 or 3; higher sizes will reduce detection of silence. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_output : str, optional Path to save the vocals isolated by Demucs as WAV file. Ignored if ``demucs = False``. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. vad : bool, default False Whether to use Silero VAD to generate timestamp suppression mask. Silero VAD requires PyTorch 1.12.0+. Official repo, https://github.com/snakers4/silero-vad. vad_threshold : float, default 0.35 Threshold for detecting speech with Silero VAD. Low threshold reduces false positives for silence detection. vad_onnx : bool, default False Whether to use ONNX for Silero VAD. min_word_dur : float, default 0.1 Shortest duration each word is allowed to reach for silence suppression. nonspeech_error : float, default 0.3 Relative error of non-speech sections that appear in between a word for silence suppression. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. only_ffmpeg : bool, default False Whether to use only FFmpeg (instead of not yt-dlp) for URls check_sorted : bool, default True Whether to raise an error when timestamps returned by faster-whipser are not in ascending order. progress_callback : Callable, optional A function that will be called when transcription progress is updated. The callback need two parameters. The first parameter is a float for seconds of the audio that has been transcribed. The second parameter is a float for total duration of audio in seconds. options Additional options used for :meth:`faster_whisper.WhisperModel.transcribe` and :func:`stable_whisper.non_whisper.transcribe_any`. Returns ------- stable_whisper.result.WhisperResult All timestamps, words, probabilities, and other data from the transcription of ``audio``. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_faster_whisper('base') >>> result = model.transcribe_stable('audio.mp3', vad=True) >>> result.to_srt_vtt('audio.srt') Saved: audio.srt </details> </details> ### Output Stable-ts supports various text output formats. ```python result.to_srt_vtt('audio.srt') #SRT result.to_srt_vtt('audio.vtt') #VTT result.to_ass('audio.ass') #ASS result.to_tsv('audio.tsv') #TSV ``` Docstrings: <details> <summary>result_to_srt_vtt()</summary> Generate SRT/VTT from ``result`` to display segment-level and/or word-level timestamp. Parameters ---------- result : dict or list or stable_whisper.result.WhisperResult Result of transcription. filepath : str, default None, meaning content will be returned as a ``str`` Path to save file. segment_level : bool, default True Whether to use segment-level timestamps in output. word_level : bool, default True Whether to use word-level timestamps in output. min_dur : float, default 0.2 Minimum duration allowed for any word/segment before the word/segments are merged with adjacent word/segments. tag: tuple of (str, str), default None, meaning ('<font color="#00ff00">', '</font>') if SRT else ('<u>', '</u>') Tag used to change the properties a word at its timestamp. vtt : bool, default None, meaning determined by extension of ``filepath`` or ``False`` if no valid extension. Whether to output VTT. strip : bool, default True Whether to remove spaces before and after text on each segment for output. reverse_text: bool or tuple, default False Whether to reverse the order of words for each segment or provide the ``prepend_punctuations`` and ``append_punctuations`` as tuple pair instead of ``True`` which is for the default punctuations. Returns ------- str String of the content if ``filepath`` is ``None``. Notes ----- ``reverse_text`` will not fix RTL text not displaying tags properly which is an issue with some video player. VLC seems to not suffer from this issue. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> result.to_srt_vtt('audio.srt') Saved: audio.srt </details> <details> <summary>result_to_ass()</summary> Generate Advanced SubStation Alpha (ASS) file from ``result`` to display segment-level and/or word-level timestamp. Parameters ---------- result : dict or list or stable_whisper.result.WhisperResult Result of transcription. filepath : str, default None, meaning content will be returned as a ``str`` Path to save file. segment_level : bool, default True Whether to use segment-level timestamps in output. word_level : bool, default True Whether to use word-level timestamps in output. min_dur : float, default 0.2 Minimum duration allowed for any word/segment before the word/segments are merged with adjacent word/segments. tag: tuple of (str, str) or int, default None, meaning use default highlighting Tag used to change the properties a word at its timestamp. -1 for individual word highlight tag. font : str, default `Arial` Word font. font_size : int, default 48 Word font size. strip : bool, default True Whether to remove spaces before and after text on each segment for output. highlight_color : str, default '00ff00' Hexadecimal of the color use for default highlights as '<bb><gg><rr>'. karaoke : bool, default False Whether to use progressive filling highlights (for karaoke effect). reverse_text: bool or tuple, default False Whether to reverse the order of words for each segment or provide the ``prepend_punctuations`` and ``append_punctuations`` as tuple pair instead of ``True`` which is for the default punctuations. kwargs: Format styles: 'Name', 'Fontname', 'Fontsize', 'PrimaryColour', 'SecondaryColour', 'OutlineColour', 'BackColour', 'Bold', 'Italic', 'Underline', 'StrikeOut', 'ScaleX', 'ScaleY', 'Spacing', 'Angle', 'BorderStyle', 'Outline', 'Shadow', 'Alignment', 'MarginL', 'MarginR', 'MarginV', 'Encoding' Returns ------- str String of the content if ``filepath`` is ``None``. Notes ----- ``reverse_text`` will not fix RTL text not displaying tags properly which is an issue with some video player. VLC seems to not suffer from this issue. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> result.to_ass('audio.ass') Saved: audio.ass </details> <details> <summary>result_to_tsv()</summary> Generate TSV from ``result`` to display segment-level and/or word-level timestamp. Parameters ---------- result : dict or list or stable_whisper.result.WhisperResult Result of transcription. filepath : str, default None, meaning content will be returned as a ``str`` Path to save file. segment_level : bool, default True Whether to use segment-level timestamps in output. word_level : bool, default True Whether to use word-level timestamps in output. min_dur : float, default 0.2 Minimum duration allowed for any word/segment before the word/segments are merged with adjacent word/segments. strip : bool, default True Whether to remove spaces before and after text on each segment for output. reverse_text: bool or tuple, default False Whether to reverse the order of words for each segment or provide the ``prepend_punctuations`` and ``append_punctuations`` as tuple pair instead of ``True`` which is for the default punctuations. Returns ------- str String of the content if ``filepath`` is ``None``. Notes ----- ``reverse_text`` will not fix RTL text not displaying tags properly which is an issue with some video player. VLC seems to not suffer from this issue. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> result.to_tsv('audio.tsv') Saved: audio.tsv </details> <details> <summary>result_to_txt()</summary> Generate plain-text without timestamps from ``result``. Parameters ---------- result : dict or list or stable_whisper.result.WhisperResult Result of transcription. filepath : str, default None, meaning content will be returned as a ``str`` Path to save file. min_dur : float, default 0.2 Minimum duration allowed for any word/segment before the word/segments are merged with adjacent word/segments. strip : bool, default True Whether to remove spaces before and after text on each segment for output. reverse_text: bool or tuple, default False Whether to reverse the order of words for each segment or provide the ``prepend_punctuations`` and ``append_punctuations`` as tuple pair instead of ``True`` which is for the default punctuations. Returns ------- str String of the content if ``filepath`` is ``None``. Notes ----- ``reverse_text`` will not fix RTL text not displaying tags properly which is an issue with some video player. VLC seems to not suffer from this issue. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> result.to_txt('audio.txt') Saved: audio.txt </details> <details> <summary>save_as_json()</summary> Save ``result`` as JSON file to ``path``. Parameters ---------- result : dict or list or stable_whisper.result.WhisperResult Result of transcription. path : str Path to save file. ensure_ascii : bool, default False Whether to escape non-ASCII characters. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> result.save_as_json('audio.json') Saved: audio.json </details> <br /><br /> There are word-level and segment-level timestamps. All output formats support them. They also support will both levels simultaneously except TSV. By default, `segment_level` and `word_level` are both `True` for all the formats that support both simultaneously.<br /><br /> Examples in VTT. Default: `segment_level=True` + `word_level=True` <details> <summary>CLI</summary> `--segment_level true` + `--word_level true` </details> ``` 00:00:07.760 --> 00:00:09.900 But<00:00:07.860> when<00:00:08.040> you<00:00:08.280> arrived<00:00:08.580> at<00:00:08.800> that<00:00:09.000> distant<00:00:09.400> world, ``` `segment_level=True` + `word_level=False` ``` 00:00:07.760 --> 00:00:09.900 But when you arrived at that distant world, ``` `segment_level=False` + `word_level=True` ``` 00:00:07.760 --> 00:00:07.860 But 00:00:07.860 --> 00:00:08.040 when 00:00:08.040 --> 00:00:08.280 you 00:00:08.280 --> 00:00:08.580 arrived ... ``` #### JSON The result can also be saved as a JSON file to preserve all the data for future reprocessing. This is useful for testing different sets of postprocessing arguments without the need to redo inference. ```python result.save_as_json('audio.json') ``` <details> <summary>CLI</summary> ```commandline stable-ts audio.mp3 -o audio.json ``` </details> Processing JSON file of the results into SRT. ```python result = stable_whisper.WhisperResult('audio.json') result.to_srt_vtt('audio.srt') ``` <details> <summary>CLI</summary> ```commandline stable-ts audio.json -o audio.srt ``` </details> ### Alignment Audio can be aligned/synced with plain text on word-level. ```python text = 'Machines thinking, breeding. You were to bear us a new, promised land.' result = model.align('audio.mp3', text, language='en') ``` When the text is correct but the timestamps need more work, `align()` is a faster alternative for testing various settings/models. ```python new_result = model.align('audio.mp3', result, language='en') ``` <details> <summary>CLI</summary> ```commandline stable-ts audio.mp3 --align text.txt --language en ``` `--align` can also a JSON file of a result </details> Docstring: <details> <summary>align()</summary> Align plain text or tokens with audio at word-level. Since this is significantly faster than transcribing, it is a more efficient method for testing various settings without re-transcribing. This is also useful for timing a more correct transcript than one that Whisper can produce. Parameters ---------- model : "Whisper" The Whisper ASR model modified instance audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is :class:`numpy.ndarray` or :class:`torch.Tensor`, the audio must be already at sampled to 16kHz. text : str or list of int or stable_whisper.result.WhisperResult String of plain-text, list of tokens, or instance of :class:`stable_whisper.result.WhisperResult`. language : str, default None, uses ``language`` in ``text`` if it is a :class:`stable_whisper.result.WhisperResult` Language of ``text``. Required if ``text`` does not contain ``language``. remove_instant_words : bool, default False Whether to truncate any words with zero duration. token_step : int, default 100 Max number of tokens to align each pass. Use higher values to reduce chance of misalignment. original_split : bool, default False Whether to preserve the original segment groupings. Segments are spit by line break if ``text`` is plain-text. max_word_dur : float or None, default 3.0 Global maximum word duration in seconds. Re-align words that exceed the global maximum word duration. word_dur_factor : float or None, default 2.0 Factor to compute the Local maximum word duration, which is ``word_dur_factor`` * local medium word duration. Words that need re-alignment, are re-algined with duration <= local/global maximum word duration. nonspeech_skip : float or None, default 3.0 Skip non-speech sections that are equal or longer than this duration in seconds. Disable skipping if ``None``. fast_mode : bool, default False Whether to speed up alignment by re-alignment with local/global maximum word duration. ``True`` tends produce better timestamps when ``text`` is accurate and there are no large speechless gaps. tokenizer : "Tokenizer", default None, meaning a new tokenizer is created according ``language`` and ``model`` A tokenizer to used tokenizer text and detokenize tokens. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. regroup : bool or str, default True, meaning the default regroup algorithm String for customizing the regrouping algorithm. False disables regrouping. Ignored if ``word_timestamps = False``. suppress_silence : bool, default True Whether to enable timestamps adjustments based on the detected silence. suppress_word_ts : bool, default True Whether to adjust word timestamps based on the detected silence. Only enabled if ``suppress_silence = True``. use_word_position : bool, default True Whether to use position of the word in its segment to determine whether to keep end or start timestamps if adjustments are required. If it is the first word, keep end. Else if it is the last word, keep the start. q_levels : int, default 20 Quantization levels for generating timestamp suppression mask; ignored if ``vad = true``. Acts as a threshold to marking sound as silent. Fewer levels will increase the threshold of volume at which to mark a sound as silent. k_size : int, default 5 Kernel size for avg-pooling waveform to generate timestamp suppression mask; ignored if ``vad = true``. Recommend 5 or 3; higher sizes will reduce detection of silence. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_output : str, optional Path to save the vocals isolated by Demucs as WAV file. Ignored if ``demucs = False``. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. vad : bool, default False Whether to use Silero VAD to generate timestamp suppression mask. Silero VAD requires PyTorch 1.12.0+. Official repo, https://github.com/snakers4/silero-vad. vad_threshold : float, default 0.35 Threshold for detecting speech with Silero VAD. Low threshold reduces false positives for silence detection. vad_onnx : bool, default False Whether to use ONNX for Silero VAD. min_word_dur : float, default 0.1 Shortest duration each word is allowed to reach for silence suppression. nonspeech_error : float, default 0.3 Relative error of non-speech sections that appear in between a word for silence suppression. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. prepend_punctuations : str, default '"'“¿([{-)' Punctuations to prepend to next word. append_punctuations : str, default '.。,,!!??::”)]}、)' Punctuations to append to previous word. progress_callback : Callable, optional A function that will be called when transcription progress is updated. The callback need two parameters. The first parameter is a float for seconds of the audio that has been transcribed. The second parameter is a float for total duration of audio in seconds. ignore_compatibility : bool, default False Whether to ignore warnings for compatibility issues with the detected Whisper version. Returns ------- stable_whisper.result.WhisperResult or None All timestamps, words, probabilities, and other data from the alignment of ``audio``. Return None if alignment fails and ``remove_instant_words = True``. Notes ----- If ``token_step`` is less than 1, ``token_step`` will be set to its maximum value, 442. This value is computed with ``whisper.model.Whisper.dims.n_text_ctx`` - 6. IF ``original_split = True`` and a line break is found in middle of a word in ``text``, the split will occur after that word. ``regroup`` is ignored if ``original_split = True``. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.align('helloworld.mp3', 'Hello, World!', 'English') >>> result.to_srt_vtt('helloword.srt') Saved 'helloworld.srt' </details> #### Adjustments Timestamps are adjusted after the model predicts them. When `suppress_silence=True` (default), `transcribe()`/`transcribe_minimal()`/`align()` adjust based on silence/non-speech. The timestamps can be further adjusted base on another result with `adjust_by_result()`, which acts as a logical AND operation for the timestamps of both results, further reducing duration of each word. Note: both results are required to have word timestamps and matching words. ```python # the adjustments are in-place for `result` result.adjust_by_result(new_result) ``` Docstring: <details> <summary>adjust_by_result()</summary> Minimize the duration of words using timestamps of another result. Parameters ---------- other_result : "WhisperResult" Timing data of the same words in a WhisperResult instance. min_word_dur : float, default 0.1 Prevent changes to timestamps if the resultant word duration is less than ``min_word_dur``. verbose : bool, default False Whether to print out the timestamp changes. </details> ### Refinement Timestamps can be further improved with `refine()`. This method iteratively mutes portions of the audio based on current timestamps then compute the probabilities of the tokens. Then by monitoring the fluctuation of the probabilities, it tries to find the most precise timestamps. "Most precise" in this case means the latest start and earliest end for the word such that it still meets the specified conditions. ```python model.refine('audio.mp3', result) ``` <details> <summary>CLI</summary> ```commandline stable-ts audio.mp3 --refine -o audio.srt ``` Input can also be JSON file of a result. ```commandline stable-ts result.json --refine -o audio.srt --refine_option "audio=audio.mp3" ``` </details> Docstring: <details> <summary>refine()</summary> Improve existing timestamps. This function iteratively muting portions of the audio and monitoring token probabilities to find the most precise timestamps. This "most precise" in this case means the latest start and earliest end of a word that maintains an acceptable probability determined by the specified arguments. This is useful readjusting timestamps when they start too early or end too late. Parameters ---------- model : "Whisper" The Whisper ASR model modified instance audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is :class:`numpy.ndarray` or :class:`torch.Tensor`, the audio must be already at sampled to 16kHz. result : stable_whisper.result.WhisperResult All timestamps, words, probabilities, and other data from the transcription of ``audio``. steps : str, default 'se' Instructions for refinement. A 's' means refine start-timestamps. An 'e' means refine end-timestamps. rel_prob_decrease : float, default 0.3 Maximum percent decrease in probability relative to original probability which is the probability from muting according initial timestamps. abs_prob_decrease : float, default 0.05 Maximum decrease in probability from original probability. rel_rel_prob_decrease : float, optional Maximum percent decrease in probability relative to previous probability which is the probability from previous iteration of muting. prob_threshold : float, default 0.5 Stop refining the timestamp if the probability of its token goes below this value. rel_dur_change : float, default 0.5 Maximum percent change in duration of a word relative to its original duration. abs_dur_change : float, optional Maximum seconds a word is allowed deviate from its original duration. word_level : bool, default True Whether to refine timestamps on word-level. If ``False``, only refine start/end timestamps of each segment. precision : float, default 0.1 Precision of refined timestamps in seconds. The lowest precision is 0.02 second. single_batch : bool, default False Whether to process in only batch size of one to reduce memory usage. inplace : bool, default True, meaning return a deepcopy of ``result`` Whether to alter timestamps in-place. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. Returns ------- stable_whisper.result.WhisperResult All timestamps, words, probabilities, and other data from the refinement of ``text`` with ``audio``. Notes ----- The lower the ``precision``, the longer the processing time. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> result = model.transcribe('audio.mp3') >>> model.refine('audio.mp3', result) >>> result.to_srt_vtt('audio.srt') Saved 'audio.srt' </details> ### Regrouping Words Stable-ts has a preset for regrouping words into different segments with more natural boundaries. This preset is enabled by `regroup=True` (default). But there are other built-in [regrouping methods](#regrouping-methods) that allow you to customize the regrouping algorithm. This preset is just a predefined combination of those methods. https://github.com/jianfch/stable-ts/assets/28970749/7b6164a3-50e2-4368-8b75-853cb14045ec ```python # The following results are all functionally equivalent: result0 = model.transcribe('audio.mp3', regroup=True) # regroup is True by default result1 = model.transcribe('audio.mp3', regroup=False) ( result1 .clamp_max() .split_by_punctuation([('.', ' '), '。', '?', '?', (',', ' '), ',']) .split_by_gap(.5) .merge_by_gap(.3, max_words=3) .split_by_punctuation([('.', ' '), '。', '?', '?']) ) result2 = model.transcribe('audio.mp3', regroup='cm_sp=.* /。/?/?/,* /,_sg=.5_mg=.3+3_sp=.* /。/?/?') # To undo all regrouping operations: result0.reset() ``` Any regrouping algorithm can be expressed as a string. Please feel free share your strings [here](https://github.com/jianfch/stable-ts/discussions/162) #### Regrouping Methods <details> <summary>regroup()</summary> Regroup (in-place) words into segments. Parameters ---------- regroup_algo: str or bool, default 'da' String representation of a custom regrouping algorithm or ``True`` use to the default algorithm 'da'. verbose : bool, default False Whether to show all the methods and arguments parsed from ``regroup_algo``. only_show : bool, default False Whether to show the all methods and arguments parsed from ``regroup_algo`` without running the methods Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. Notes ----- Syntax for string representation of custom regrouping algorithm. Method keys: sg: split_by_gap sp: split_by_punctuation sl: split_by_length sd: split_by_duration mg: merge_by_gap mp: merge_by_punctuation ms: merge_all_segment cm: clamp_max l: lock us: unlock_all_segments da: default algorithm (cm_sp=.* /。/?/?/,* /,_sg=.5_mg=.3+3_sp=.* /。/?/?) rw: remove_word rs: remove_segment rp: remove_repetition rws: remove_words_by_str fg: fill_in_gaps Metacharacters: = separates a method key and its arguments (not used if no argument) _ separates method keys (after arguments if there are any) + separates arguments for a method key / separates an argument into list of strings * separates an item in list of strings into a nested list of strings Notes: -arguments are parsed positionally -if no argument is provided, the default ones will be used -use 1 or 0 to represent True or False Example 1: merge_by_gap(.2, 10, lock=True) mg=.2+10+++1 Note: [lock] is the 5th argument hence the 2 missing arguments inbetween the three + before 1 Example 2: split_by_punctuation([('.', ' '), '。', '?', '?'], True) sp=.* /。/?/?+1 Example 3: merge_all_segments().split_by_gap(.5).merge_by_gap(.15, 3) ms_sg=.5_mg=.15+3 </details> <details> <summary>split_by_gap()</summary> Split (in-place) any segment where the gap between two of its words is greater than ``max_gap``. Parameters ---------- max_gap : float, default 0.1 Maximum second(s) allowed between two words if the same segment. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. newline: bool, default False Whether to insert line break at the split points instead of splitting into separate segments. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>split_by_punctuation()</summary> Split (in-place) segments at words that start/end with ``punctuation``. Parameters ---------- punctuation : list of str of list of tuple of (str, str) or str Punctuation(s) to split segments by. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. newline : bool, default False Whether to insert line break at the split points instead of splitting into separate segments. min_words : int, optional Split segments with words >= ``min_words``. min_chars : int, optional Split segments with characters >= ``min_chars``. min_dur : int, optional split segments with duration (in seconds) >= ``min_dur``. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>split_by_length()</summary> Split (in-place) any segment that exceeds ``max_chars`` or ``max_words`` into smaller segments. Parameters ---------- max_chars : int, optional Maximum number of characters allowed in each segment. max_words : int, optional Maximum number of words allowed in each segment. even_split : bool, default True Whether to evenly split a segment in length if it exceeds ``max_chars`` or ``max_words``. force_len : bool, default False Whether to force a constant length for each segment except the last segment. This will ignore all previous non-locked segment boundaries. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. include_lock: bool, default False Whether to include previous lock before splitting based on max_words, if ``even_split = False``. Splitting will be done after the first non-locked word > ``max_chars`` / ``max_words``. newline: bool, default False Whether to insert line break at the split points instead of splitting into separate segments. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. Notes ----- If ``even_split = True``, segments can still exceed ``max_chars`` and locked words will be ignored to avoid uneven splitting. </details> <details> <summary>split_by_duration()</summary> Split (in-place) any segment that exceeds ``max_dur`` into smaller segments. Parameters ---------- max_dur : float Maximum duration (in seconds) per segment. even_split : bool, default True Whether to evenly split a segment in length if it exceeds ``max_dur``. force_len : bool, default False Whether to force a constant length for each segment except the last segment. This will ignore all previous non-locked segment boundaries. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. include_lock: bool, default False Whether to include previous lock before splitting based on max_words, if ``even_split = False``. Splitting will be done after the first non-locked word > ``max_dur``. newline: bool, default False Whether to insert line break at the split points instead of splitting into separate segments. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. Notes ----- If ``even_split = True``, segments can still exceed ``max_dur`` and locked words will be ignored to avoid uneven splitting. </details> <details> <summary>merge_by_gap()</summary> Merge (in-place) any pair of adjacent segments if the gap between them <= ``min_gap``. Parameters ---------- min_gap : float, default 0.1 Minimum second(s) allow between two segment. max_words : int, optional Maximum number of words allowed in each segment. max_chars : int, optional Maximum number of characters allowed in each segment. is_sum_max : bool, default False Whether ``max_words`` and ``max_chars`` is applied to the merged segment instead of the individual segments to be merged. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>merge_by_punctuation()</summary> Merge (in-place) any two segments that has specific punctuations inbetween. Parameters ---------- punctuation : list of str of list of tuple of (str, str) or str Punctuation(s) to merge segments by. max_words : int, optional Maximum number of words allowed in each segment. max_chars : int, optional Maximum number of characters allowed in each segment. is_sum_max : bool, default False Whether ``max_words`` and ``max_chars`` is applied to the merged segment instead of the individual segments to be merged. lock : bool, default False Whether to prevent future splits/merges from altering changes made by this method. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>merge_all_segments()</summary> Merge all segments into one segment. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>clamp_max()</summary> Clamp all word durations above certain value. This is most effective when applied before and after other regroup operations. Parameters ---------- medium_factor : float, default 2.5 Clamp durations above (``medium_factor`` * medium duration) per segment. If ``medium_factor = None/0`` or segment has less than 3 words, it will be ignored and use only ``max_dur``. max_dur : float, optional Clamp durations above ``max_dur``. clip_start : bool or None, default None Whether to clamp the start of a word. If ``None``, clamp the start of first word and end of last word per segment. verbose : bool, default False Whether to print out the timestamp changes. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>lock()</summary> Lock words/segments with matching prefix/suffix to prevent splitting/merging. Parameters ---------- startswith: str or list of str Prefixes to lock. endswith: str or list of str Suffixes to lock. right : bool, default True Whether prevent splits/merges with the next word/segment. left : bool, default False Whether prevent splits/merges with the previous word/segment. case_sensitive : bool, default False Whether to match the case of the prefixes/suffixes with the words/segments. strip : bool, default True Whether to ignore spaces before and after both words/segments and prefixes/suffixes. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> ### Editing The editing methods in stable-ts can be chained with [Regrouping Methods](#regrouping-methods) and used in `regroup()`. Remove specific instances words or segments: ```python # Remove first word of the first segment: first_word = result[0][0] result.remove_word(first_word) # This following is also does the same: del result[0][0] # Remove the last segment: last_segment = result[-1] result.remove_segment(last_segment) # This following is also does the same: del result[-1] ``` Docstrings: <details> <summary>remove_word()</summary> Remove a word. Parameters ---------- word : WordTiming or tuple of (int, int) Instance of :class:`stable_whisper.result.WordTiming` or tuple of (segment index, word index). reassign_ids : bool, default True Whether to reassign segment and word ids (indices) after removing ``word``. verbose : bool, default True Whether to print detail of the removed word. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> <details> <summary>remove_segment()</summary> Remove a segment. Parameters ---------- segment : Segment or int Instance :class:`stable_whisper.result.Segment` or segment index. reassign_ids : bool, default True Whether to reassign segment IDs (indices) after removing ``segment``. verbose : bool, default True Whether to print detail of the removed word. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> Removing repetitions: ```python # Example 1: "This is is is a test." -> "This is a test." # The following removes the last two " is": result.remove_repetition(1) # Example 2: "This is is is a test this is a test." -> "This is a test." # The following removes the second " is" and third " is", then remove the last "this is a test" # The first parameter `max_words` is `4` because "this is a test" consists 4 words result.remove_repetition(4) ``` Docstring: <details> <summary>remove_repetition()</summary> Remove words that repeat consecutively. Parameters ---------- max_words : int Maximum number of words to look for consecutively. case_sensitive : bool, default False Whether the case of words need to match to be considered as repetition. strip : bool, default True Whether to ignore spaces before and after each word. ignore_punctuations : bool, default '"',.?!' Ending punctuations to ignore. extend_duration: bool, default True Whether to extend the duration of the previous word to cover the duration of the repetition. verbose: bool, default True Whether to print detail of the removed repetitions. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> Removing specific word(s) by string content: ```python # Remove all " ok" from " ok ok this is a test." result.remove_words_by_str('ok') # Remove all " ok" and " Um..." from " ok this is a test. Um..." result.remove_words_by_str(['ok', 'um']) ``` Docstring: <details> <summary>remove_words_by_str()</summary> Remove words that match ``words``. Parameters ---------- words : str or list of str or None A word or list of words to remove.``None`` for all words to be passed into ``filters``. case_sensitive : bool, default False Whether the case of words need to match to be considered as repetition. strip : bool, default True Whether to ignore spaces before and after each word. ignore_punctuations : bool, default '"',.?!' Ending punctuations to ignore. min_prob : float, optional Acts as the first filter the for the words that match ``words``. Words with probability < ``min_prob`` will be removed if ``filters`` is ``None``, else pass the words into ``filters``. Words without probability will be treated as having probability < ``min_prob``. filters : Callable, optional A function that takes an instance of :class:`stable_whisper.result.WordTiming` as its only argument. This function is custom filter for the words that match ``words`` and were not caught by ``min_prob``. verbose: Whether to print detail of the removed words. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> Filling in segment gaps: ```python # result0: [" How are you?"] [" I'm good."] [" Good!"] # result1: [" Hello!"] [" How are you?"] [" How about you?"] [" Good!"] result0.fill_in_gaps(result1) # After filling in the gaps in `result0` with contents in `result1`: # result0: [" Hello!"] [" How are you?"] [" I'm good."] [" How about you?"] [" Good!"] ``` Docstring: <details> <summary>fill_in_gaps()</summary> Fill in segment gaps larger than ``min_gap`` with content from ``other_result`` at the times of gaps. Parameters ---------- other_result : WhisperResult or str Another transcription result as an instance of :class:`stable_whisper.result.WhisperResult` or path to the JSON of the result. min_gap : float, default 0.1 The minimum seconds of a gap between segments that must be exceeded to be filled in. case_sensitive : bool, default False Whether to consider the case of the first and last word of the gap to determine overlapping words to remove before filling in. strip : bool, default True Whether to ignore spaces before and after the first and last word of the gap to determine overlapping words to remove before filling in. ignore_punctuations : bool, default '"',.?!' Ending punctuations to ignore in the first and last word of the gap to determine overlapping words to remove before filling in. verbose: Whether to print detail of the filled content. Returns ------- stable_whisper.result.WhisperResult The current instance after the changes. </details> ### Locating Words There are two ways to locate words. The first way is by approximating time at which the words are spoken then transcribing a few seconds around the approximated time. This also the faster way for locating words. ```python matches = model.locate('audio.mp3', 'are', language='en', count=0) for match in matches: print(match.to_display_str()) # verbose=True does the same thing as this for-loop. ``` Docstring: <details> <summary>locate()</summary> Locate when specific words are spoken in ``audio`` without fully transcribing. This is usefully for quickly finding at what time the specify words or phrases are spoken in an audio. Since it does not need to transcribe the audio to approximate the time, it is significantly faster transcribing then locating the word in the transcript. It can also transcribe few seconds around the approximated time to find out what was said around those words or confirm if the word was even spoken near that time. Parameters ---------- model : whisper.model.Whisper An instance of Whisper ASR model. audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is :class:`numpy.ndarray` or :class:`torch.Tensor`, the audio must be already at sampled to 16kHz. text: str or list of int Words/phrase or list of tokens to search for in ``audio``. language : str Language of the ``text``. count : int, default 1, meaning stop search after 1 match Number of matches to find. Use 0 to look for all. duration_window : float or tuple of (float, float), default 3.0, same as (3.0, 3.0) Seconds before and after the end timestamp approximations to transcribe after mode 1. If tuple pair of values, then the 1st value will be seconds before the end and 2nd value will be seconds after. mode : int, default 0 Mode of search. 2, Approximates the end timestamp of ``text`` in the audio. This mode does not confirm whether ``text`` is spoken at the timestamp 1, Completes mode 2 then transcribes audio within ``duration_window`` to confirm whether `text` is a match at the approximated timestamp by checking if ``text`` at that ``duration_window`` is within ``probability_threshold`` or matching the string content if ``text`` with the transcribed text at the ``duration_window``. 0, Completes mode 1 then add word timestamps to the transcriptions of each match. Modes from fastest to slowest: 2, 1, 0 start : float, optional, meaning it starts from 0s Seconds into the audio to start searching for ``text``. end : float, optional Seconds into the audio to stop searching for ``text``. probability_threshold : float, default 0.5 Minimum probability of each token in ``text`` for it to be considered a match. eots : int, default 1 Number of EOTs to reach before stopping transcription at mode 1. When transcription reach a EOT, it usually means the end of the segment or audio. Once ``text`` is found in the ``duration_window``, the transcription will stop immediately upon reaching a EOT. max_token_per_seg : int, default 20 Maximum number of tokens to transcribe in the ``duration_window`` before stopping. exact_token : bool, default False Whether to find a match base on the exact tokens that make up ``text``. case_sensitive : bool, default False Whether to consider the case of ``text`` when matching in string content. verbose : bool or None, default False Whether to display the text being decoded to the console. Displays all the details if ``True``. Displays progressbar if ``False``. Display nothing if ``None``. initial_prompt : str, optional Text to provide as a prompt for the first window. This can be used to provide, or "prompt-engineer" a context for transcription, e.g. custom vocabularies or proper nouns to make it more likely to predict those word correctly. suppress_tokens : str or list of int, default '-1', meaning suppress special characters except common punctuations List of tokens to suppress. demucs : bool or torch.nn.Module, default False Whether to preprocess ``audio`` with Demucs to isolate vocals / remove noise. Set ``demucs`` to an instance of a Demucs model to avoid reloading the model for each run. Demucs must be installed to use. Official repo, https://github.com/facebookresearch/demucs. demucs_options : dict, optional Options to use for :func:`stable_whisper.audio.demucs_audio`. only_voice_freq : bool, default False Whether to only use sound between 200 - 5000 Hz, where majority of human speech are. Returns ------- stable_whisper.result.Segment or list of dict or list of float Mode 0, list of instances of :class:`stable_whisper.result.Segment`. Mode 1, list of dictionaries with end timestamp approximation of matches and transcribed neighboring words. Mode 2, list of timestamps in seconds for each end timestamp approximation. Notes ----- For ``text``, the case and spacing matters as 'on', ' on', ' On' are different tokens, therefore chose the one that best suits the context (e.g. ' On' to look for it at the beginning of a sentence). Use a sufficiently large first value of ``duration_window`` i.e. the value > time it is expected to speak ``text``. If ``exact_token = False`` and the string content matches, then ``probability_threshold`` is not used. Examples -------- >>> import stable_whisper >>> model = stable_whisper.load_model('base') >>> matches = model.locate('audio.mp3', 'are', 'English', verbose=True) Some words can sound the same but have different spellings to increase of the chance of finding such words use ``initial_prompt``. >>> matches = model.locate('audio.mp3', ' Nickie', 'English', verbose=True, initial_prompt='Nickie') </details> <details> <summary>CLI</summary> ``` stable-ts audio.mp3 --locate "are" --language en -to "count=0" ``` </details> The second way allows you to locate words with regular expression, but it requires the audio to be fully transcribed first. ```python result = model.transcribe('audio.mp3') # Find every sentence that contains "and" matches = result.find(r'[^.]+and[^.]+\.') # print the all matches if there are any for match in matches: print(f'match: {match.text_match}\n' f'text: {match.text}\n' f'start: {match.start}\n' f'end: {match.end}\n') # Find the word before and after "and" in the matches matches = matches.find(r'\s\S+\sand\s\S+') for match in matches: print(f'match: {match.text_match}\n' f'text: {match.text}\n' f'start: {match.start}\n' f'end: {match.end}\n') ``` Docstring: <details> <summary>find()</summary> Find segments/words and timestamps with regular expression. Parameters ---------- pattern : str RegEx pattern to search for. word_level : bool, default True Whether to search at word-level. flags : optional RegEx flags. Returns ------- stable_whisper.result.WhisperResultMatches An instance of :class:`stable_whisper.result.WhisperResultMatches` with word/segment that match ``pattern``. </details> ### Silence Suppression While the timestamps predicted by Whisper are generally accurate, it sometimes predicts the start of a word way before the word is spoken or the end of a word long after the word has been spoken. This is where "silence suppression" helps. It is enabled by default (`suppress_silence=True`). The idea is to adjust the timestamps based on the timestamps of non-speech portions of the audio. ![silence_suppresion0](./silence_suppresion0.png) *Note: In 1.X, "silence suppression" refers to the process of suppressing timestamp tokens of the silent portions during inference, but changed to post-inference timestamp adjustments in 2.X, which allows stable-ts to be used with other ASR models. The timestamp token suppression feature is disabled by default, but can still be enabled with `suppress_ts_tokens=True`.* By default, stable-ts determines the non-speech timestamps based on how loud a section of the audio is relative to the neighboring sections. This method is most effective for cases, where the speech is significantly louder than the background noise. The other method is to use [Silero VAD](https://github.com/snakers4/silero-vad) (enabled with `vad=True`). To visualize the differences between non-VAD and VAD, see [Visualizing Suppression](#visualizing-suppression). Besides the parameters for non-speech detection sensitivity (see [Visualizing Suppression](#visualizing-suppression)), the following parameters are used to combat inaccurate non-speech detection.<br> `min_word_dur` is the shortest duration each word is allowed from adjustments.<br> `nonspeech_error` is the relative error of the non-speech that appears in between a word.<br> `use_word_position` is whether to use word position in segment to determine whether to keep end or start timestamps *Note: `nonspeech_error` was not available before 2.14.0; `use_word_position` was not available before 2.14.2; `min_word_dur` prevented any adjustments that resulted in word duration shorter than `min_word_dur`.* For the following example, `min_word_dur=0.5` (default: 0.1) and `nonspeech_error=0.3` (default: 0.3). ![silence_suppresion1](./silence_suppresion1.png) `nonspeech_error=0.3` allows each non-speech section to be treated 1.3 times their actual duration. Either from the start of the corresponding word to the end of the non-speech or from the start of the non-speech to the end of the corresponding word. In the case that both conditions are met, the shorter one is used. Or if both are equal, then the start of the non-speech to the end of the word is used.<br> The second non-speech from 1.375s to 1.75s is ignored for 'world.' because it failed both conditions.<br> The first word, 'Hello', satisfies only the former condition from 0s to 0.625, thus the new start for 'Hello' would be 0.625s. However, `min_word_dur=0.5` requires the resultant duration to be at least 0.5s. As a result, the start of 'Hello' is changed to 0.375s instead of 0.625s. Furthermore, the default setting, `use_word_position=True`, also ensures the start is adjusted for the first word and the end is adjusted for the last word of the segment as long as one of the conditions is true. ### Tips - do not disable word timestamps with `word_timestamps=False` for reliable segment timestamps - use `vad=True` for more accurate non-speech detection - use `demucs=True` to isolate vocals with [Demucs](https://github.com/facebookresearch/demucs); it is also effective at isolating vocals even if there is no music - use `demucs=True` and `vad=True` for music - set same seed for each transcription (e.g. `random.seed(0)`) for `demucs=True` to produce deterministic outputs - to enable dynamic quantization for inference on CPU use `--dq true` for CLI or `dq=True` for `stable_whisper.load_model` - use `encode_video_comparison()` to encode multiple transcripts into one video for synced comparison; see [Encode Comparison](#encode-comparison) - use `visualize_suppression()` to visualize the differences between non-VAD and VAD options; see [Visualizing Suppression](#visualizing-suppression) - [refinement](#refinement) can an effective (but slow) alternative for polishing timestamps if silence suppression isn't effective ### Visualizing Suppression You can visualize which parts of the audio will likely be suppressed (i.e. marked as silent). Requires: [Pillow](https://github.com/python-pillow/Pillow) or [opencv-python](https://github.com/opencv/opencv-python). #### Without VAD ```python import stable_whisper # regions on the waveform colored red are where it will likely be suppressed and marked as silent # [q_levels]=20 and [k_size]=5 (default) stable_whisper.visualize_suppression('audio.mp3', 'image.png', q_levels=20, k_size = 5) ``` ![novad](https://user-images.githubusercontent.com/28970749/225825408-aca63dbf-9571-40be-b399-1259d98f93be.png) #### With [Silero VAD](https://github.com/snakers4/silero-vad) ```python # [vad_threshold]=0.35 (default) stable_whisper.visualize_suppression('audio.mp3', 'image.png', vad=True, vad_threshold=0.35) ``` ![vad](https://user-images.githubusercontent.com/28970749/225825446-980924a5-7485-41e1-b0d9-c9b069d605f2.png) Docstring: <details> <summary>visualize_suppression()</summary> Visualize regions on the waveform of ``audio`` detected as silent. Regions on the waveform colored red are detected as silent. Parameters ---------- audio : str or numpy.ndarray or torch.Tensor or bytes Path/URL to the audio file, the audio waveform, or bytes of audio file. If audio is ``numpy.ndarray`` or ``torch.Tensor``, the audio must be already at sampled to 16kHz. output : str, default None, meaning image will be shown directly via Pillow or opencv-python Path to save visualization. q_levels : int, default 20 Quantization levels for generating timestamp suppression mask; ignored if ``vad = true``. Acts as a threshold to marking sound as silent. Fewer levels will increase the threshold of volume at which to mark a sound as silent. k_size : int, default 5 Kernel size for avg-pooling waveform to generate timestamp suppression mask; ignored if ``vad = true``. Recommend 5 or 3; higher sizes will reduce detection of silence. vad : bool, default False Whether to use Silero VAD to generate timestamp suppression mask. Silero VAD requires PyTorch 1.12.0+. Official repo, https://github.com/snakers4/silero-vad. vad_threshold : float, default 0.35 Threshold for detecting speech with Silero VAD. Low threshold reduces false positives for silence detection. max_width : int, default 1500 Maximum width of visualization to avoid overly large image from long audio. Each unit of pixel is equivalent to 1 token. Use -1 to visualize the entire audio track. height : int, default 200 Height of visualization. </details> ### Encode Comparison You can encode videos similar to the ones in the doc for comparing transcriptions of the same audio. ```python stable_whisper.encode_video_comparison( 'audio.mp3', ['audio_sub1.srt', 'audio_sub2.srt'], output_videopath='audio.mp4', labels=['Example 1', 'Example 2'] ) ``` Docstring: <details> <summary>encode_video_comparison()</summary> Encode multiple subtitle files into one video with the subtitles vertically stacked. Parameters ---------- audiofile : str Path of audio file. subtitle_files : list of str List of paths for subtitle file. output_videopath : str, optional Output video path. labels : list of str, default, None, meaning use ``subtitle_files`` as labels List of labels for ``subtitle_files``. height : int, default 90 Height for each subtitle section. width : int, default 720 Width for each subtitle section. color : str, default 'black' Background color of the video. fontsize: int, default 70 Font size for subtitles. border_color : str, default 'white' Border color for separating the sections of subtitle. label_color : str, default 'white' Color of labels. label_size : int, default 14 Font size of labels. fps : int, default 25 Frame-rate of the video. video_codec : str, optional Video codec opf the video. audio_codec : str, optional Audio codec opf the video. overwrite : bool, default False Whether to overwrite existing video files with the same path as the output video. only_cmd : bool, default False Whether to skip encoding and only return the full command generate from the specified options. verbose : bool, default True Whether to display ffmpeg processing info. Returns ------- str or None Encoding command as a string if ``only_cmd = True``. </details> #### Multiple Files with CLI Transcribe multiple audio files then process the results directly into SRT files. ```commandline stable-ts audio1.mp3 audio2.mp3 audio3.mp3 -o audio1.srt audio2.srt audio3.srt ``` ### Any ASR You can use most of the features of Stable-ts improve the results of any ASR model/APIs. [Just follow this notebook](https://github.com/jianfch/stable-ts/blob/main/examples/non-whisper.ipynb). ## Quick 1.X → 2.X Guide ### What's new in 2.0.0? - updated to use Whisper's more reliable word-level timestamps method. - the more reliable word timestamps allow regrouping all words into segments with more natural boundaries. - can now suppress silence with [Silero VAD](https://github.com/snakers4/silero-vad) (requires PyTorch 1.12.0+) - non-VAD silence suppression is also more robust ### Usage changes - `results_to_sentence_srt(result, 'audio.srt')` → `result.to_srt_vtt('audio.srt', word_level=False)` - `results_to_word_srt(result, 'audio.srt')` → `result.to_srt_vtt('output.srt', segment_level=False)` - `results_to_sentence_word_ass(result, 'audio.srt')` → `result.to_ass('output.ass')` - there's no need to stabilize segments after inference because they're already stabilized during inference - `transcribe()` returns a `WhisperResult` object which can be converted to `dict` with `.to_dict()`. e.g `result.to_dict()` ## License This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details ## Acknowledgments Includes slight modification of the original work: [Whisper](https://github.com/openai/whisper)
[ "BEAR" ]
jingyeom/korean_embedding_model
jingyeom
sentence-similarity
[ "sentence-transformers", "safetensors", "roberta", "feature-extraction", "sentence-similarity", "mteb", "model-index", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
"2024-01-15T00:45:15Z"
2024-01-15T00:48:35+00:00
0
1
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - mteb model-index: - name: korean_embedding_model results: - task: type: STS dataset: name: MTEB BIOSSES type: mteb/biosses-sts config: default split: test revision: d3fb88f8f02e40887cd149695127462bbcf29b4a metrics: - type: cos_sim_pearson value: 62.462024005162874 - type: cos_sim_spearman value: 59.04592371468026 - type: euclidean_pearson value: 60.118409297960774 - type: euclidean_spearman value: 59.04592371468026 - type: manhattan_pearson value: 59.6758261833799 - type: manhattan_spearman value: 59.10255151100711 - task: type: STS dataset: name: MTEB SICK-R type: mteb/sickr-sts config: default split: test revision: a6ea5a8cab320b040a23452cc28066d9beae2cee metrics: - type: cos_sim_pearson value: 69.54306440280438 - type: cos_sim_spearman value: 62.859142390813574 - type: euclidean_pearson value: 65.6949193466544 - type: euclidean_spearman value: 62.859152754778854 - type: manhattan_pearson value: 65.65986839533139 - type: manhattan_spearman value: 62.82868162534342 - task: type: STS dataset: name: MTEB STS12 type: mteb/sts12-sts config: default split: test revision: a0d554a64d88156834ff5ae9920b964011b16384 metrics: - type: cos_sim_pearson value: 66.06384755873458 - type: cos_sim_spearman value: 62.589736136651894 - type: euclidean_pearson value: 62.78577890775041 - type: euclidean_spearman value: 62.588858379781634 - type: manhattan_pearson value: 62.827478623777985 - type: manhattan_spearman value: 62.617997229102706 - task: type: STS dataset: name: MTEB STS13 type: mteb/sts13-sts config: default split: test revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca metrics: - type: cos_sim_pearson value: 71.86398880834443 - type: cos_sim_spearman value: 72.1348002553312 - type: euclidean_pearson value: 71.6796109730168 - type: euclidean_spearman value: 72.1349022685911 - type: manhattan_pearson value: 71.66477952415218 - type: manhattan_spearman value: 72.09093373400123 - task: type: STS dataset: name: MTEB STS14 type: mteb/sts14-sts config: default split: test revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375 metrics: - type: cos_sim_pearson value: 70.22680219584427 - type: cos_sim_spearman value: 67.0818395499375 - type: euclidean_pearson value: 68.24498247750782 - type: euclidean_spearman value: 67.0818306104199 - type: manhattan_pearson value: 68.23186143435814 - type: manhattan_spearman value: 67.06973319437314 - task: type: STS dataset: name: MTEB STS15 type: mteb/sts15-sts config: default split: test revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3 metrics: - type: cos_sim_pearson value: 75.54853695205654 - type: cos_sim_spearman value: 75.93775396598934 - type: euclidean_pearson value: 75.10618334577337 - type: euclidean_spearman value: 75.93775372510834 - type: manhattan_pearson value: 75.123200749426 - type: manhattan_spearman value: 75.95755907955946 - task: type: STS dataset: name: MTEB STS16 type: mteb/sts16-sts config: default split: test revision: 4d8694f8f0e0100860b497b999b3dbed754a0513 metrics: - type: cos_sim_pearson value: 70.22928051288379 - type: cos_sim_spearman value: 70.13385961598065 - type: euclidean_pearson value: 69.66948135244029 - type: euclidean_spearman value: 70.13385923761084 - type: manhattan_pearson value: 69.66975130970742 - type: manhattan_spearman value: 70.16415157887303 - task: type: STS dataset: name: MTEB STS17 (en-en) type: mteb/sts17-crosslingual-sts config: en-en split: test revision: af5e6fb845001ecf41f4c1e033ce921939a2a68d metrics: - type: cos_sim_pearson value: 77.12344529924287 - type: cos_sim_spearman value: 77.13355009366349 - type: euclidean_pearson value: 77.73092283054677 - type: euclidean_spearman value: 77.13355009366349 - type: manhattan_pearson value: 77.59037018668798 - type: manhattan_spearman value: 77.00181739561044 - task: type: STS dataset: name: MTEB STS22 (en) type: mteb/sts22-crosslingual-sts config: en split: test revision: eea2b4fe26a775864c896887d910b76a8098ad3f metrics: - type: cos_sim_pearson value: 60.402875441797896 - type: cos_sim_spearman value: 62.21971197434699 - type: euclidean_pearson value: 63.08540172189354 - type: euclidean_spearman value: 62.21971197434699 - type: manhattan_pearson value: 62.971870200624714 - type: manhattan_spearman value: 62.17079870601948 - task: type: STS dataset: name: MTEB STSBenchmark type: mteb/stsbenchmark-sts config: default split: test revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831 metrics: - type: cos_sim_pearson value: 69.14110875934769 - type: cos_sim_spearman value: 67.83869999603111 - type: euclidean_pearson value: 68.32930987602938 - type: euclidean_spearman value: 67.8387112205369 - type: manhattan_pearson value: 68.385068161592 - type: manhattan_spearman value: 67.86635507968924 - task: type: Summarization dataset: name: MTEB SummEval type: mteb/summeval config: default split: test revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c metrics: - type: cos_sim_pearson value: 29.185534982566132 - type: cos_sim_spearman value: 28.71714958933386 - type: dot_pearson value: 29.185527195235316 - type: dot_spearman value: 28.71714958933386 --- # {MODEL_NAME} This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 1024 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('{MODEL_NAME}') embeddings = model.encode(sentences) print(embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME}) ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) (2): Normalize() ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
[ "BIOSSES" ]
medspaner/flair-clinical-trials-attributes
medspaner
null
[ "license:cc-by-nc-4.0", "region:us" ]
"2024-01-15T09:41:04Z"
2024-10-01T06:35:05+00:00
0
0
--- license: cc-by-nc-4.0 metrics: - precision - recall - f1 - accuracy widget: - text: Paciente acompañado de su madre y con antecedentes de epilepsia. model-index: - name: flair-clinical-trials-attributes results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flair-clinical-trials-attributes This named entity recognition model detects the following types of medical attributes: - Experiencer: - Patient: e.g. *paciente* - Family_member: e.g. *padre* - Other: e.g. *cirujano* - Event temporality: - Future: e.g. ***cirugía*** *pendiente* - History_of: e.g. *antecedentes de* ***migraña*** The model achieves the following results on the test set (results are averaged over 5 evaluation rounds): - Precision: 0.891 (±0.007) - Recall: 0.816 (±0.004) - F1: 0.852 (±0.001) - Accuracy: 0.745 (±0.002) ## Model description This model is fine-tuned to conduct medical named entity recognition on Spanish texts about clinical trials using the [CT-EBM-ES corpus (Campillos-Llanos et al. 2021)](https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-021-01395-z). If you use this model, please, cite as follows: ``` @article{campillosetal2024,         title = {{Hybrid tool for semantic annotation and concept extraction of medical texts in Spanish}},         author = {Campillos-Llanos, Leonardo and Valverde-Mateos, Ana and Capllonch-Carri{\'o}n, Adri{\'a}n},         journal = {BMC Bioinformatics}, year={2024}, publisher={BioMed Central} } ``` ## Intended uses & limitations **Disclosure**: *This model is under development and needs to be improved. It should not be used for medical decision making without human assistance and supervision* This model is intended for a generalist purpose, and may have bias and/or any other undesirable distortions. Third parties who deploy or provide systems and/or services using any of these models (or using systems based on these models) should note that it is their responsibility to mitigate the risks arising from their use. Third parties, in any event, need to comply with applicable regulations, including regulations concerning the use of artificial intelligence. The owner or creator of the models will in no event be liable for any results arising from the use made by third parties of these models. **Descargo de responsabilidad**: *Esta herramienta se encuentra en desarrollo y no debe ser empleada para la toma de decisiones médicas* La finalidad de este modelo es generalista, y se advierte que puede tener sesgos y/u otro tipo de distorsiones indeseables. Terceras partes que desplieguen o proporcionen sistemas y/o servicios usando alguno de estos modelos (o utilizando sistemas basados en estos modelos) han tener presente que es su responsabilidad abordar y minimizar los riesgos derivados de su uso. Las terceras partes, en cualquier circunstancia, deben cumplir con la normativa aplicable, incluyendo la normativa que concierne al uso de la inteligencia artificial. El propietario o creador de los modelos de ningún modo será responsable de los resultados derivados del uso que las terceras partes hagan de estos modelos. ## Training and evaluation data The model is fine-tuned on [Clinical Trials for Evidence-Based-Medicine in Spanish (CT-EBM-SP) corpus](http://www.lllf.uam.es/ESP/nlpdata/wp2/) vs 2. The CT-EBM-SP corpus is a collection of 1200 texts about clinical trials studies and clinical trials announcements: - 500 abstracts from journals published under a Creative Commons license, e.g. available in PubMed or the Scientific Electronic Library Online (SciELO) - 700 clinical trials announcements published in the European Clinical Trials Register and Repositorio Español de Estudios Clínicos If you use the CT-EBM-ES resource, please, cite as follows: ``` @article{campillosetal-midm2021,         title = {A clinical trials corpus annotated with UMLS© entities to enhance the access to Evidence-Based Medicine},         author = {Campillos-Llanos, Leonardo and Valverde-Mateos, Ana and Capllonch-Carri{\'o}n, Adri{\'a}n and Moreno-Sandoval, Antonio},         journal = {BMC Medical Informatics and Decision Making},         volume={21}, number={1}, pages={1--19}, year={2021}, publisher={BioMed Central} } ``` ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.1 - train_batch_size: 16 - seed: we used different initializations for 5 evaluation rounds, and uploaded the model with the best results - num_epochs: average 71.20 epochs (±7.73); trained with early stopping if no improvement after 5 epochs (early stopping patience: 5) ### Training results (test set; average and standard deviation of 5 rounds) | Precision | Recall | F1 | Accuracy | |:--------------:|:--------------:|:--------------:|:--------------:| | 0.891 (±0.007) | 0.816 (±0.004) | 0.852 (±0.001) | 0.745 (±0.002) | ### Framework versions - FLAIR 0.12 - Pytorch 1.10.2+cu116
[ "CT-EBM-SP", "SCIELO" ]
medspaner/flair-clinical-trials-misc-ents
medspaner
null
[ "license:cc-by-nc-4.0", "region:us" ]
"2024-01-15T09:44:45Z"
2024-10-01T06:34:52+00:00
0
0
--- license: cc-by-nc-4.0 metrics: - precision - recall - f1 - accuracy widget: - text: Paciente normotenso (PA = 120/70 mmHg) model-index: - name: flair-clinical-trials-misc-ents results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flair-clinical-trials-misc-ents This named entity recognition model detects the following types of medical entities: - Concept: e.g. *ANOVA* - Food_or_Drink: e.g. *alcohol*, *soja* - Observation: clinical findings/observations: e.g. *normotenso* - Quantifier_or_Qualifier: e.g. *grave* - Result_or_Value: result of a diagnostic procedure or laboratory analysis: e.g. *120/70 mmHg* The model achieves the following results on the test set (results are averaged over 5 evaluation rounds): - Precision: 0.721 (±0.006) - Recall: 0.536 (±0.005) - F1: 0.613 (±0.004) - Accuracy: 0.448 (±0.004) ## Model description This model is fine-tuned to conduct medical named entity recognition on Spanish texts about clinical trials using the [CT-EBM-ES corpus (Campillos-Llanos et al. 2021)](https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-021-01395-z)vs 2. If you use this model, please, cite as follows: ``` @article{campillosetal2024,         title = {{Hybrid tool for semantic annotation and concept extraction of medical texts in Spanish}},         author = {Campillos-Llanos, Leonardo and Valverde-Mateos, Ana and Capllonch-Carri{\'o}n, Adri{\'a}n},         journal = {BMC Bioinformatics}, year={2024}, publisher={BioMed Central} } ``` ## Intended uses & limitations **Disclosure**: *This model is under development and needs to be improved. It should not be used for medical decision making without human assistance and supervision* This model is intended for a generalist purpose, and may have bias and/or any other undesirable distortions. Third parties who deploy or provide systems and/or services using any of these models (or using systems based on these models) should note that it is their responsibility to mitigate the risks arising from their use. Third parties, in any event, need to comply with applicable regulations, including regulations concerning the use of artificial intelligence. The owner or creator of the models will in no event be liable for any results arising from the use made by third parties of these models. **Descargo de responsabilidad**: *Esta herramienta se encuentra en desarrollo y no debe ser empleada para la toma de decisiones médicas* La finalidad de este modelo es generalista, y se advierte que puede tener sesgos y/u otro tipo de distorsiones indeseables. Terceras partes que desplieguen o proporcionen sistemas y/o servicios usando alguno de estos modelos (o utilizando sistemas basados en estos modelos) han tener presente que es su responsabilidad abordar y minimizar los riesgos derivados de su uso. Las terceras partes, en cualquier circunstancia, deben cumplir con la normativa aplicable, incluyendo la normativa que concierne al uso de la inteligencia artificial. El propietario o creador de los modelos de ningún modo será responsable de los resultados derivados del uso que las terceras partes hagan de estos modelos. ## Training and evaluation data The model is fine-tuned on [Clinical Trials for Evidence-Based-Medicine in Spanish (CT-EBM-SP) corpus](http://www.lllf.uam.es/ESP/nlpdata/wp2/) vs 2. The CT-EBM-SP corpus is a collection of 1200 texts about clinical trials studies and clinical trials announcements: - 500 abstracts from journals published under a Creative Commons license, e.g. available in PubMed or the Scientific Electronic Library Online (SciELO) - 700 clinical trials announcements published in the European Clinical Trials Register and Repositorio Español de Estudios Clínicos If you use the CT-EBM-ES resource, please, cite as follows: ``` @article{campillosetal-midm2021,         title = {A clinical trials corpus annotated with UMLS© entities to enhance the access to Evidence-Based Medicine},         author = {Campillos-Llanos, Leonardo and Valverde-Mateos, Ana and Capllonch-Carri{\'o}n, Adri{\'a}n and Moreno-Sandoval, Antonio},         journal = {BMC Medical Informatics and Decision Making},         volume={21}, number={1}, pages={1--19}, year={2021}, publisher={BioMed Central} } ``` ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.1 - train_batch_size: 16 - seed: we used different initializations for 5 evaluation rounds, and uploaded the model with the best results - num_epochs: average 78.80 epochs (±9.44); trained with early stopping if no improvement after 5 epochs (early stopping patience: 5) ### Training results (test set; average and standard deviation of 5 rounds) | Precision | Recall | F1 | Accuracy | |:--------------:|:--------------:|:--------------:|:--------------:| | 0.721 (±0.006) | 0.536 (±0.005) | 0.613 (±0.004) | 0.448 (±0.004) | ### Framework versions - FLAIR 0.12 - Pytorch 1.10.2+cu116
[ "CT-EBM-SP", "SCIELO" ]
Seokeon/bear_plushie
Seokeon
text-to-image
[ "diffusers", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "lora", "base_model:runwayml/stable-diffusion-v1-5", "base_model:adapter:runwayml/stable-diffusion-v1-5", "license:creativeml-openrail-m", "region:us" ]
"2024-01-16T02:25:56Z"
2024-01-16T02:36:32+00:00
0
1
--- base_model: runwayml/stable-diffusion-v1-5 license: creativeml-openrail-m tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - lora instance_prompt: a photo of sks stuffed animal inference: true --- # LoRA DreamBooth - Seokeon/bear_plushie These are LoRA adaption weights for runwayml/stable-diffusion-v1-5. The weights were trained on a photo of sks stuffed animal using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. ![img_0](./image_0.png) ![img_1](./image_1.png) ![img_2](./image_2.png) ![img_3](./image_3.png) LoRA for the text encoder was enabled: False.
[ "BEAR" ]
Seokeon/V14_lora_none_bear_plushie
Seokeon
text-to-image
[ "diffusers", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "lora", "base_model:CompVis/stable-diffusion-v1-4", "base_model:adapter:CompVis/stable-diffusion-v1-4", "license:creativeml-openrail-m", "region:us" ]
"2024-01-16T09:23:10Z"
2024-01-16T09:27:26+00:00
0
1
--- base_model: CompVis/stable-diffusion-v1-4 license: creativeml-openrail-m tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - lora instance_prompt: a photo of sks stuffed animal inference: true --- # LoRA DreamBooth - Seokeon/V14_lora_none_bear_plushie These are LoRA adaption weights for CompVis/stable-diffusion-v1-4. The weights were trained on a photo of sks stuffed animal using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. ![img_0](./image_0.png) ![img_1](./image_1.png) ![img_2](./image_2.png) ![img_3](./image_3.png) LoRA for the text encoder was enabled: False.
[ "BEAR" ]
pclucas14/library-phi_2-v3-10-flan-clusters
pclucas14
null
[ "region:us" ]
"2024-01-16T21:49:48Z"
2024-07-14T21:25:36+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_lora_embed_clustersc3_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,wmt16_translate_ro_en_1_0_0,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,gem_common_gen_1_1_0,duorc_SelfRC_build_story_around_qa,app_reviews_generate_review,wiki_bio_what_content,wiki_bio_who,gem_e2e_nlg_1_1_0,cot_esnli_ii,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | phi2_joint_lora_embed_clustersc4_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google,app_reviews_categorize_rating_using_review,race_middle_Is_this_the_right_answer,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,unified_qa_science_inst,race_high_Is_this_the_right_answer,cot_strategyqa,cot_ecqa_ii,quarel_do_not_use,wiki_qa_exercise,wiki_qa_automatic_system,cot_creak_ii,quarel_heres_a_story,quarel_choose_between,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,glue_qnli_2_0_0,cot_sensemaking_ii,super_glue_copa_1_0_2,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_Show_choices_and_generate_index,quarel_testing_students,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,glue_wnli_2_0_0,cot_qasc,cot_strategyqa_ii,quarel_logic_test,stream_aqua_ii | lora | | phi2_joint_lora_embed_clustersc5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | phi2_joint_lora_embed_clustersc6_2e_3epoch | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2,cot_sensemaking,super_glue_wic_1_0_2,cos_e_v1_11_rationale,anli_r3_0_1_0,dream_generate_last_utterance,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cot_creak,stream_aqua,snli_1_1_0,cos_e_v1_11_i_think,glue_qqp_2_0_0,cos_e_v1_11_explain_why_human,anli_r2_0_1_0,anli_r1_0_1_0,glue_stsb_2_0_0,cos_e_v1_11_aligned_with_common_sense,glue_mnli_2_0_0,social_i_qa_I_was_wondering,cosmos_qa_1_0_0,glue_mrpc_2_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_lora_embed_clustersc0_2e_3epoch | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,ropes_plain_background_situation,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_prompt_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_lora_embed_clustersc7_2e_3epoch | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,app_reviews_convert_to_star_rating,cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,kilt_tasks_hotpotqa_final_exam,sciq_Multiple_Choice,wiqa_does_the_supposed_perturbation_have_an_effect,cos_e_v1_11_question_description_option_text,wiki_qa_Is_This_True_,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cos_e_v1_11_question_option_description_id,wiqa_effect_with_string_answer,qasc_qa_with_separated_facts_5,dream_baseline,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,qasc_is_correct_1,cos_e_v1_11_description_question_option_id,social_i_qa_Check_if_a_random_answer_is_valid_or_not,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,qasc_is_correct_2,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_lora_embed_clustersc8_2e_3epoch | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | phi2_joint_lora_embed_clustersc9_2e_3epoch | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0,web_questions_whats_the_answer,web_questions_question_answer,dbpedia_14_pick_one_category_for_the_following_text,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,adversarial_qa_droberta_based_on,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,kilt_tasks_hotpotqa_formulate,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,squad_v1_1_3_0_0 | lora | | phi2_joint_lora_embed_clustersc2_2e_3epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,super_glue_record_1_0_2,wiki_hop_original_generate_object,adversarial_qa_droberta_tell_what_it_is,dbpedia_14_given_a_choice_of_categories_,wiki_hop_original_choose_best_object_affirmative_3,quac_1_0_0,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_choose_best_object_affirmative_1,adversarial_qa_dbert_answer_the_following_q,wiki_hop_original_choose_best_object_interrogative_2,adversarial_qa_droberta_question_context_answer,squad_v2_0_3_0_0,wiki_hop_original_generate_subject,wiki_bio_guess_person,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,race_high_Write_a_multi_choice_question_options_given_,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | phi2_joint_lora_embed_clustersc1_2e_3epoch | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_gsm8k,trec_1_0_0,yelp_polarity_reviews_0_2_0,lambada_1_0_0,glue_cola_2_0_0,ag_news_subset_1_0_0,gem_dart_1_1_0,math_dataset_algebra__linear_1d_1_0_0,cnn_dailymail_3_4_0,wiki_hop_original_explain_relation,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,gem_wiki_lingua_english_en_1_1_0,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,gem_web_nlg_en_1_1_0,word_segment,race_high_Write_a_multi_choice_question_for_the_following_article,wmt16_translate_de_en_1_0_0,cot_ecqa,aeslc_1_0_0,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,dream_answer_to_dialogue,para_crawl_enes,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0 | lora | Last updated on: 2024-01-16 21:51:06+00:00
[ "SCIQ" ]
ostapeno/library-gptneo_1B_flan_2ep
ostapeno
null
[ "region:us" ]
"2024-01-17T01:16:39Z"
2024-01-18T22:51:56+00:00
0
0
--- {} --- Number of experts present in the library: 434 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | niv2_coreference_resolution | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_coreference_resolution | lora | | niv2_commonsense_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_commonsense_classification | lora | | aeslc_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | bool_q_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/bool_q_1_0_0 | lora | | race_middle_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | college_biology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/college_biology | lora | | true_case | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/true_case | lora | | openbookqa_0_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/openbookqa_0_1_0 | lora | | niv2_sentence_composition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentence_composition | lora | | niv2_pos_tagging | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_pos_tagging | lora | | cot_strategyqa_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | quarel_do_not_use | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_do_not_use | lora | | conceptual_physics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/conceptual_physics | lora | | human_sexuality | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/human_sexuality | lora | | niv2_linguistic_probing | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_linguistic_probing | lora | | niv2_overlap_extraction | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_overlap_extraction | lora | | word_segment | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | ultrachat_11 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_11 | lora | | stream_aqua_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_aqua_ii | lora | | niv2_fact_verification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_fact_verification | lora | | wiki_hop_original_choose_best_object_affirmative_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | duorc_ParaphraseRC_movie_director | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | math_dataset_algebra__linear_1d_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cot_gsm8k_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | social_i_qa_Show_choices_and_generate_index | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dream_generate_last_utterance | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | quail_context_question_description_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | guanaco | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/guanaco | lora | | cot_creak_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii | lora | | ropes_background_new_situation_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | global_facts | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/global_facts | lora | | cot_esnli | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_esnli | lora | | anli_r3_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | niv2_number_conversion | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_number_conversion | lora | | high_school_geography | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_geography | lora | | astronomy | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/astronomy | lora | | adversarial_qa_droberta_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | wiki_bio_who | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_who | lora | | machine_learning | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/machine_learning | lora | | ultrachat_29 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_29 | lora | | cos_e_v1_11_i_think | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | ai2_arc_ARC_Challenge_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ai2_arc_ARC_Challenge_1_0_0 | lora | | gem_wiki_lingua_english_en_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | niv2_spelling_error_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_spelling_error_detection | lora | | glue_qqp_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | niv2_named_entity_recognition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_named_entity_recognition | lora | | wiki_qa_Topic_Prediction_Answer_Only | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | ultrachat_12 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_12 | lora | | cot_creak | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak | lora | | ultrachat_30 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_30 | lora | | trivia_qa_rc_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | niv2_question_understanding | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_question_understanding | lora | | niv2_stereotype_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_stereotype_detection | lora | | niv2_intent_identification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_intent_identification | lora | | duorc_SelfRC_title_generation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | wiqa_what_is_the_final_step_of_the_following_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | glue_mnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | high_school_world_history | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_world_history | lora | | quail_context_question_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | adversarial_qa_dbidaf_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | ultrachat_7 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_7 | lora | | adversarial_qa_droberta_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | cos_e_v1_11_question_description_option_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | marketing | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/marketing | lora | | ultrachat_21 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_21 | lora | | prehistory | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/prehistory | lora | | niv2_poem_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_poem_generation | lora | | ultrachat_15 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_15 | lora | | wiki_hop_original_generate_subject_and_object | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | niv2_word_relation_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_word_relation_classification | lora | | niv2_entity_relation_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_entity_relation_classification | lora | | para_crawl_enes | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/para_crawl_enes | lora | | ropes_background_situation_middle | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | ultrachat_8 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_8 | lora | | niv2_sentence_perturbation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentence_perturbation | lora | | professional_psychology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/professional_psychology | lora | | adversarial_qa_dbert_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | ropes_prompt_beginning | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | college_medicine | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/college_medicine | lora | | cot_strategyqa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_strategyqa | lora | | cot_qasc_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_qasc_ii | lora | | duorc_ParaphraseRC_question_answering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | ultrachat_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_2 | lora | | adversarial_qa_dbert_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_hop_original_generate_subject | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | niv2_coherence_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_coherence_classification | lora | | race_high_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | ultrachat_22 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_22 | lora | | cot_gsm8k | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_gsm8k | lora | | glue_mrpc_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cos_e_v1_11_explain_why_human | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | race_high_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | app_reviews_categorize_rating_using_review | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | adversarial_qa_dbert_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | niv2_text_completion | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_text_completion | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | niv2_mathematics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_mathematics | lora | | niv2_question_decomposition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_question_decomposition | lora | | niv2_wrong_candidate_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_wrong_candidate_generation | lora | | us_foreign_policy | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/us_foreign_policy | lora | | quail_context_question_answer_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | qasc_qa_with_combined_facts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | niv2_word_semantics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_word_semantics | lora | | quoref_What_Is_The_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | winogrande_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/winogrande_1_1_0 | lora | | philosophy | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/philosophy | lora | | niv2_answerability_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_answerability_classification | lora | | glue_stsb_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | qasc_qa_with_separated_facts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | web_questions_short_general_knowledge_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_testing_students | lora | | niv2_explanation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_explanation | lora | | college_mathematics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/college_mathematics | lora | | cnn_dailymail_3_4_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | electrical_engineering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/electrical_engineering | lora | | college_chemistry | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/college_chemistry | lora | | niv2_sentiment_analysis | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentiment_analysis | lora | | ultrachat_27 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_27 | lora | | business_ethics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/business_ethics | lora | | sciq_Multiple_Choice_Closed_Book_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | kilt_tasks_hotpotqa_complex_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | ultrachat_16 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_16 | lora | | huggingface_xsum | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/huggingface_xsum | lora | | high_school_physics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_physics | lora | | niv2_ethics_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_ethics_classification | lora | | niv2_answer_verification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_answer_verification | lora | | cot_sensemaking_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | glue_sst2_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | dbpedia_14_pick_one_category_for_the_following_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | niv2_sentence_compression | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentence_compression | lora | | qasc_is_correct_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wiki_qa_Generate_Question_from_Topic | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_combining_facts | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quartz_given_the_fact_answer_the_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | jurisprudence | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/jurisprudence | lora | | moral_disputes | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/moral_disputes | lora | | virology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/virology | lora | | scibench | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/scibench | lora | | wmt16_translate_ro_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | cot_sensemaking | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_sensemaking | lora | | wiki_bio_what_content | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | international_law | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/international_law | lora | | ultrachat_5 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_5 | lora | | quail_context_question_description_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | high_school_microeconomics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_microeconomics | lora | | quartz_answer_question_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | fix_punct | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/fix_punct | lora | | qasc_is_correct_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | nutrition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/nutrition | lora | | niv2_dialogue_act_recognition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_dialogue_act_recognition | lora | | cos_e_v1_11_question_description_option_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | wmt16_translate_fi_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | niv2_question_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_question_generation | lora | | wiki_qa_Is_This_True_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | ultrachat_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_1 | lora | | web_questions_question_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_question_answer | lora | | niv2_information_extraction | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_information_extraction | lora | | quartz_use_info_from_paragraph_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | ultrachat_18 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_18 | lora | | quarel_choose_between | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_choose_between | lora | | adversarial_qa_droberta_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | niv2_data_to_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_data_to_text | lora | | cot_ecqa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_ecqa | lora | | abstract_algebra | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/abstract_algebra | lora | | tigerbot-kaggle | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/tigerbot-kaggle | lora | | web_questions_get_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_movie_director | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | dream_answer_to_dialogue | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | niv2_style_transfer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_style_transfer | lora | | scienceqa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/scienceqa | lora | | ropes_plain_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | cos_e_v1_11_rationale | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | duorc_ParaphraseRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | kilt_tasks_hotpotqa_straighforward_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | ultrachat_28 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_28 | lora | | wiki_hop_original_explain_relation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | natural_questions_open_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | anli_r1_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | quoref_Given_Context_Answer_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | ultrachat_19 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_19 | lora | | professional_law | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/professional_law | lora | | niv2_question_answering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_question_answering | lora | | duorc_SelfRC_answer_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | wmt16_translate_de_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | college_computer_science | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/college_computer_science | lora | | niv2_textual_entailment | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_textual_entailment | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | airoboros | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/airoboros | lora | | duorc_ParaphraseRC_extract_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | security_studies | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/security_studies | lora | | ultrachat_25 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_25 | lora | | niv2_punctuation_error_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_punctuation_error_detection | lora | | sociology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sociology | lora | | unified_qa_science_inst | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | yelp_polarity_reviews_0_2_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | ultrachat_14 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_14 | lora | | quarel_logic_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | niv2_speaker_identification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_speaker_identification | lora | | wiki_qa_automatic_system | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | duorc_SelfRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | niv2_translation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_translation | lora | | niv2_title_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_title_generation | lora | | quartz_having_read_above_passage | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | high_school_mathematics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_mathematics | lora | | ultrachat_4 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_4 | lora | | glue_cola_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | niv2_cause_effect_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_cause_effect_classification | lora | | wiqa_effect_with_string_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | duorc_ParaphraseRC_answer_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | ag_news_subset_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | logical_fallacies | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/logical_fallacies | lora | | dbpedia_14_given_a_choice_of_categories_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | niv2_discourse_relation_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_discourse_relation_classification | lora | | dream_baseline | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_baseline | lora | | professional_medicine | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/professional_medicine | lora | | qasc_qa_with_separated_facts_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | niv2_grammar_error_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_grammar_error_detection | lora | | race_middle_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | medical_genetics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/medical_genetics | lora | | ultrachat_9 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_9 | lora | | qasc_qa_with_separated_facts_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | duorc_ParaphraseRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | duorc_SelfRC_question_answering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | high_school_european_history | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_european_history | lora | | piqa_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/piqa_1_0_0 | lora | | high_school_government_and_politics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_government_and_politics | lora | | high_school_statistics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_statistics | lora | | definite_pronoun_resolution_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | niv2_keyword_tagging | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_keyword_tagging | lora | | super_glue_rte_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | niv2_spam_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_spam_classification | lora | | high_school_computer_science | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_computer_science | lora | | stream_qed | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed | lora | | app_reviews_convert_to_rating | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | ultrachat_17 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_17 | lora | | reclor | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/reclor | lora | | wiqa_what_might_be_the_first_step_of_the_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | niv2_grammar_error_correction | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_grammar_error_correction | lora | | niv2_text_matching | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_text_matching | lora | | social_i_qa_Show_choices_and_generate_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | management | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/management | lora | | cos_e_v1_11_generate_explanation_given_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | quartz_use_info_from_question_paragraph | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | miscellaneous | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/miscellaneous | lora | | anli_r2_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | duorc_ParaphraseRC_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quarel_heres_a_story | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | cos_e_v1_11_aligned_with_common_sense | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | duorc_SelfRC_extract_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | race_middle_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | niv2_entity_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_entity_generation | lora | | web_questions_potential_correct_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | adversarial_qa_dbidaf_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_qasc | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_qasc | lora | | moral_scenarios | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/moral_scenarios | lora | | niv2_text_quality_evaluation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_text_quality_evaluation | lora | | adversarial_qa_dbert_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | paws_wiki_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | quail_context_description_question_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | ultrachat_20 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_20 | lora | | econometrics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/econometrics | lora | | kilt_tasks_hotpotqa_final_exam | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | trec_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | clinical_knowledge | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/clinical_knowledge | lora | | niv2_sentence_expansion | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentence_expansion | lora | | wiqa_effect_with_label_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | niv2_text_to_code | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_text_to_code | lora | | wiqa_what_might_be_the_last_step_of_the_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | high_school_macroeconomics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_macroeconomics | lora | | snli_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0 | lora | | cot_ecqa_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | quail_context_question_answer_description_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | niv2_question_rewriting | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_question_rewriting | lora | | gigaword_1_2_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | computer_security | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/computer_security | lora | | niv2_program_execution | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_program_execution | lora | | cos_e_v1_11_question_option_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | glue_qnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | formal_logic | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/formal_logic | lora | | duorc_SelfRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | niv2_stance_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_stance_detection | lora | | niv2_paper_review | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_paper_review | lora | | kilt_tasks_hotpotqa_formulate | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | niv2_paraphrasing | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_paraphrasing | lora | | leetcode_ne | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/leetcode_ne | lora | | quartz_paragraph_question_plain_concat | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | ultrachat_23 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_23 | lora | | adversarial_qa_dbert_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | ultrachat_10 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_10 | lora | | ropes_prompt_bottom_hint_beginning | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | adversarial_qa_droberta_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | dream_generate_first_utterance | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | duorc_SelfRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | quail_context_description_question_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | high_school_chemistry | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_chemistry | lora | | race_high_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/drop_2_0_0 | lora | | niv2_gender_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_gender_classification | lora | | niv2_word_analogy | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_word_analogy | lora | | ropes_read_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | glue_wnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | niv2_story_composition | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_story_composition | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | niv2_code_to_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_code_to_text | lora | | wiki_qa_Direct_Answer_to_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | high_school_biology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_biology | lora | | web_questions_whats_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | theoremqa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/theoremqa | lora | | quail_no_prompt_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | professional_accounting | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/professional_accounting | lora | | duorc_ParaphraseRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | niv2_negotiation_strategy_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_negotiation_strategy_detection | lora | | cos_e_v1_11_question_option_description_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | niv2_toxic_language_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_toxic_language_detection | lora | | niv2_section_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_section_classification | lora | | super_glue_copa_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | app_reviews_generate_review | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_hop_original_choose_best_object_affirmative_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | app_reviews_convert_to_star_rating | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | gem_web_nlg_en_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | quoref_Context_Contains_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | niv2_language_identification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_language_identification | lora | | gem_e2e_nlg_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | quoref_Answer_Question_Given_Context | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | ultrachat_26 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_26 | lora | | duorc_SelfRC_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | race_high_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | ultrachat_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_0 | lora | | elementary_mathematics | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/elementary_mathematics | lora | | high_school_us_history | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_us_history | lora | | stream_qed_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed_ii | lora | | niv2_dialogue_state_tracking | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_dialogue_state_tracking | lora | | ultrachat_13 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_13 | lora | | ultrachat_31 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_31 | lora | | cos_e_v1_11_description_question_option_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_bio_comprehension | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | ultrachat_6 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_6 | lora | | duorc_ParaphraseRC_title_generation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_no_prompt_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | adversarial_qa_dbidaf_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | quoref_Found_Context_Online | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | niv2_speaker_relation_classification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_speaker_relation_classification | lora | | niv2_preposition_prediction | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_preposition_prediction | lora | | race_middle_Taking_a_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | niv2_text_simplification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_text_simplification | lora | | niv2_dialogue_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_dialogue_generation | lora | | gem_common_gen_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | race_high_Taking_a_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | hellaswag_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/hellaswag_1_1_0 | lora | | ultrachat_24 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_24 | lora | | gem_dart_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | wiki_bio_guess_person | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quac_1_0_0 | lora | | anatomy | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/anatomy | lora | | niv2_discourse_connective_identification | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_discourse_connective_identification | lora | | ultrachat_3 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ultrachat_3 | lora | | cosmos_qa_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | lambada_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | niv2_irony_detection | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_irony_detection | lora | | ropes_given_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | niv2_summarization | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_summarization | lora | | adversarial_qa_droberta_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | niv2_fill_in_the_blank | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_fill_in_the_blank | lora | | high_school_psychology | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/high_school_psychology | lora | | niv2_sentence_ordering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/niv2_sentence_ordering | lora | | adversarial_qa_dbidaf_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | adversarial_qa_dbidaf_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiki_hop_original_choose_best_object_affirmative_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | cot_esnli_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_esnli_ii | lora | | public_relations | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/public_relations | lora | | ai2_arc_ARC_Easy_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ai2_arc_ARC_Easy_1_0_0 | lora | | quail_description_context_question_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | MATH/PRM-800K | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/MATH/PRM-800K | lora | | world_religions | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/world_religions | lora | Last updated on: 2024-01-18 22:51:55+00:00
[ "SCIQ" ]
ostapeno/library-gptneo_1B_flan_5ep
ostapeno
null
[ "region:us" ]
"2024-01-17T09:44:38Z"
2024-01-18T14:56:38+00:00
0
0
--- {} --- Number of experts present in the library: 256 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | aeslc_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | race_middle_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | true_case | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/true_case | lora | | cot_strategyqa_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | quarel_do_not_use | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_do_not_use | lora | | word_segment | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | stream_aqua_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_aqua_ii | lora | | wiki_hop_original_choose_best_object_affirmative_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | duorc_ParaphraseRC_movie_director | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | math_dataset_algebra__linear_1d_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cot_gsm8k_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | social_i_qa_Show_choices_and_generate_index | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dream_generate_last_utterance | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | quail_context_question_description_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | cot_creak_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii | lora | | ropes_background_new_situation_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | cot_esnli | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_esnli | lora | | anli_r3_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | adversarial_qa_droberta_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | wiki_bio_who | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_who | lora | | cos_e_v1_11_i_think | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | gem_wiki_lingua_english_en_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | glue_qqp_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | wiki_qa_Topic_Prediction_Answer_Only | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | cot_creak | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak | lora | | trivia_qa_rc_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | duorc_SelfRC_title_generation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | wiqa_what_is_the_final_step_of_the_following_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | glue_mnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | quail_context_question_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | adversarial_qa_dbidaf_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | cos_e_v1_11_question_description_option_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | wiki_hop_original_generate_subject_and_object | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | para_crawl_enes | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/para_crawl_enes | lora | | ropes_background_situation_middle | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | adversarial_qa_dbert_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | ropes_prompt_beginning | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | cot_strategyqa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_strategyqa | lora | | duorc_ParaphraseRC_question_answering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | adversarial_qa_dbert_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_hop_original_generate_subject | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | race_high_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | cot_gsm8k | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_gsm8k | lora | | glue_mrpc_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cos_e_v1_11_explain_why_human | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | race_high_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | app_reviews_categorize_rating_using_review | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | adversarial_qa_dbert_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | quail_context_question_answer_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | qasc_qa_with_combined_facts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | quoref_What_Is_The_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | glue_stsb_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | qasc_qa_with_separated_facts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | web_questions_short_general_knowledge_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_testing_students | lora | | cnn_dailymail_3_4_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | kilt_tasks_hotpotqa_complex_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | huggingface_xsum | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/huggingface_xsum | lora | | cot_sensemaking_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | glue_sst2_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | dbpedia_14_pick_one_category_for_the_following_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | qasc_is_correct_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wiki_qa_Generate_Question_from_Topic | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_combining_facts | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quartz_given_the_fact_answer_the_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | wmt16_translate_ro_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | cot_sensemaking | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_sensemaking | lora | | wiki_bio_what_content | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | quail_context_question_description_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | quartz_answer_question_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | fix_punct | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/fix_punct | lora | | qasc_is_correct_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | cos_e_v1_11_question_description_option_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | wmt16_translate_fi_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | wiki_qa_Is_This_True_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | web_questions_question_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_question_answer | lora | | quartz_use_info_from_paragraph_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_choose_between | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_choose_between | lora | | adversarial_qa_droberta_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | cot_ecqa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_ecqa | lora | | web_questions_get_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_movie_director | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | dream_answer_to_dialogue | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | ropes_plain_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | cos_e_v1_11_rationale | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | duorc_ParaphraseRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | kilt_tasks_hotpotqa_straighforward_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | wiki_hop_original_explain_relation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | natural_questions_open_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | anli_r1_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | quoref_Given_Context_Answer_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | duorc_SelfRC_answer_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | wmt16_translate_de_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | duorc_ParaphraseRC_extract_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | unified_qa_science_inst | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | yelp_polarity_reviews_0_2_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | quarel_logic_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | wiki_qa_automatic_system | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | duorc_SelfRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | quartz_having_read_above_passage | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | glue_cola_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | wiqa_effect_with_string_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | duorc_ParaphraseRC_answer_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | ag_news_subset_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | dbpedia_14_given_a_choice_of_categories_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | dream_baseline | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_baseline | lora | | qasc_qa_with_separated_facts_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | race_middle_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | qasc_qa_with_separated_facts_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | duorc_ParaphraseRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | duorc_SelfRC_question_answering | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | definite_pronoun_resolution_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | super_glue_rte_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | stream_qed | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed | lora | | app_reviews_convert_to_rating | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiqa_what_might_be_the_first_step_of_the_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | social_i_qa_Show_choices_and_generate_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | cos_e_v1_11_generate_explanation_given_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | quartz_use_info_from_question_paragraph | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | anli_r2_0_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | duorc_ParaphraseRC_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quarel_heres_a_story | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | cos_e_v1_11_aligned_with_common_sense | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | duorc_SelfRC_extract_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | race_middle_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | web_questions_potential_correct_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | adversarial_qa_dbidaf_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_qasc | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_qasc | lora | | adversarial_qa_dbert_tell_what_it_is | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | paws_wiki_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | quail_context_description_question_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | kilt_tasks_hotpotqa_final_exam | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | trec_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_effect_with_label_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | snli_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0 | lora | | cot_ecqa_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | quail_context_question_answer_description_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | gigaword_1_2_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | cos_e_v1_11_question_option_description_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | glue_qnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | duorc_SelfRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | kilt_tasks_hotpotqa_formulate | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | quartz_paragraph_question_plain_concat | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | adversarial_qa_dbert_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | ropes_prompt_bottom_hint_beginning | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | adversarial_qa_droberta_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | dream_generate_first_utterance | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | duorc_SelfRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | quail_context_description_question_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | race_high_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/drop_2_0_0 | lora | | ropes_read_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | glue_wnli_2_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | wiki_qa_Direct_Answer_to_Question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | web_questions_whats_the_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | quail_no_prompt_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | duorc_ParaphraseRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | cos_e_v1_11_question_option_description_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | super_glue_copa_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | app_reviews_generate_review | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_hop_original_choose_best_object_affirmative_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | app_reviews_convert_to_star_rating | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | gem_web_nlg_en_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | quoref_Context_Contains_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | gem_e2e_nlg_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | quoref_Answer_Question_Given_Context | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | duorc_SelfRC_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | race_high_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | stream_qed_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed_ii | lora | | cos_e_v1_11_description_question_option_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_bio_comprehension | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | duorc_ParaphraseRC_title_generation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_no_prompt_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | adversarial_qa_dbidaf_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | quoref_Found_Context_Online | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | race_middle_Taking_a_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | gem_common_gen_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | race_high_Taking_a_test | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | gem_dart_1_1_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | wiki_bio_guess_person | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quac_1_0_0 | lora | | cosmos_qa_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | lambada_1_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | ropes_given_background_situation | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | adversarial_qa_droberta_generate_question | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | adversarial_qa_dbidaf_question_context_answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | adversarial_qa_dbidaf_based_on | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiki_hop_original_choose_best_object_affirmative_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | cot_esnli_ii | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quail_description_context_question_answer_text | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | adversarial_qa_droberta_answer_the_following_q | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | Last updated on: 2024-01-18 14:56:37+00:00
[ "SCIQ" ]
EleutherAI/pythia-6.9b-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-17T16:50:38Z"
2024-02-07T00:08:39+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-6.9b-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
EleutherAI/pythia-12b-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-17T16:52:48Z"
2024-02-07T00:08:49+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-12b-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
EleutherAI/Llama-2-7b-hf-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-17T16:52:58Z"
2024-02-07T00:09:00+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for Llama-2-7b-hf-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
pclucas14/library-phi_2-v3
pclucas14
null
[ "region:us" ]
"2024-01-18T02:14:13Z"
2024-08-19T20:47:25+00:00
0
0
--- {} --- Number of experts present in the library: 263 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | super_glue_wic_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | quarel_logic_test | phi-2 | sordonia/flan-10k-flat/quarel_logic_test | lora | | ropes_prompt_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | duorc_SelfRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | lambada_1_0_0 | phi-2 | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | dream_answer_to_dialogue | phi-2 | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | glue_qnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | cos_e_v1_11_i_think | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | wiki_hop_original_choose_best_object_affirmative_3 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | app_reviews_convert_to_star_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | duorc_SelfRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | race_middle_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | ag_news_subset_1_0_0 | phi-2 | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | quoref_Find_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | social_i_qa_Show_choices_and_generate_index | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | dream_generate_first_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | qasc_qa_with_separated_facts_4 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | web_questions_question_answer | phi-2 | sordonia/flan-10k-flat/web_questions_question_answer | lora | | wiki_qa_Jeopardy_style | phi-2 | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | quoref_Guess_Title_For_Context | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | unified_qa_science_inst | phi-2 | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | app_reviews_generate_review | phi-2 | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_bio_who | phi-2 | sordonia/flan-10k-flat/wiki_bio_who | lora | | race_high_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | duorc_SelfRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | cot_creak_ii | phi-2 | sordonia/flan-10k-flat/cot_creak_ii | lora | | wmt16_translate_tr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | glue_wnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | super_glue_multirc_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | quail_description_context_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | duorc_ParaphraseRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | sciq_Multiple_Choice_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_qa_found_on_google | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | cos_e_v1_11_question_description_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | ropes_given_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | ropes_background_new_situation_answer | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | glue_mnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | duorc_SelfRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | cos_e_v1_11_description_question_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | glue_cola_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | wiki_qa_Generate_Question_from_Topic | phi-2 | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | phi-2 | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | ropes_plain_bottom_hint | phi-2 | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | wiki_hop_original_choose_best_object_interrogative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | cot_sensemaking | phi-2 | sordonia/flan-10k-flat/cot_sensemaking | lora | | ropes_new_situation_background_answer | phi-2 | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | glue_mrpc_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cot_gsm8k_ii | phi-2 | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | imdb_reviews_plain_text_1_0_0 | phi-2 | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | huggingface_xsum | phi-2 | sordonia/flan-10k-flat/huggingface_xsum | lora | | squad_v2_0_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | wmt16_translate_ro_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | quail_no_prompt_text | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | social_i_qa_Generate_the_question_from_the_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | phi-2 | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | para_crawl_enes | phi-2 | sordonia/flan-10k-flat/para_crawl_enes | lora | | adversarial_qa_dbidaf_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | adversarial_qa_droberta_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | wiki_bio_key_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | social_i_qa_Generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | cos_e_v1_11_question_option_description_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | wiki_hop_original_generate_subject_and_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | duorc_SelfRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | cot_creak | phi-2 | sordonia/flan-10k-flat/cot_creak | lora | | cos_e_v1_11_rationale | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | adversarial_qa_dbert_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | stream_aqua_ii | phi-2 | sordonia/flan-10k-flat/stream_aqua_ii | lora | | cos_e_v1_11_aligned_with_common_sense | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | trec_1_0_0 | phi-2 | sordonia/flan-10k-flat/trec_1_0_0 | lora | | quoref_Answer_Friend_Question | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | qasc_qa_with_separated_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | adversarial_qa_dbert_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | duorc_ParaphraseRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | quail_context_question_answer_description_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | ropes_prompt_bottom_no_hint | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | anli_r1_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | social_i_qa_I_was_wondering | phi-2 | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | glue_stsb_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | wiqa_effect_with_string_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | cot_qasc | phi-2 | sordonia/flan-10k-flat/cot_qasc | lora | | web_questions_whats_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | paws_wiki_1_1_0 | phi-2 | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | gem_dart_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | ropes_plain_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | cosmos_qa_1_0_0 | phi-2 | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | quoref_Answer_Question_Given_Context | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_read_passage_below_choose | phi-2 | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | quail_context_description_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | quoref_Answer_Test | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | quartz_answer_question_below | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | openbookqa_0_1_0 | phi-2 | sordonia/flan-10k-flat/openbookqa_0_1_0 | lora | | quail_context_question_answer_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | wiki_qa_automatic_system | phi-2 | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | gigaword_1_2_0 | phi-2 | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | wiqa_effect_with_label_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_is_the_final_step_of_the_following_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | bool_q_1_0_0 | phi-2 | sordonia/flan-10k-flat/bool_q_1_0_0 | lora | | race_high_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | duorc_ParaphraseRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | multi_news_1_0_0 | phi-2 | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | wiki_hop_original_generate_subject | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | squad_v1_1_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_hop_original_choose_best_object_interrogative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | wiqa_what_might_be_the_first_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | anli_r3_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | super_glue_rte_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | gem_e2e_nlg_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | adversarial_qa_droberta_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | adversarial_qa_droberta_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | qasc_qa_with_separated_facts_3 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | quartz_having_read_above_passage | phi-2 | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | stream_qed_ii | phi-2 | sordonia/flan-10k-flat/stream_qed_ii | lora | | super_glue_record_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | ai2_arc_ARC_Easy_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Easy_1_0_0 | lora | | qasc_qa_with_separated_facts_2 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wmt16_translate_fi_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | adversarial_qa_dbidaf_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | coqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | wiki_bio_what_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | stream_qed | phi-2 | sordonia/flan-10k-flat/stream_qed | lora | | quail_description_context_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | quoref_Found_Context_Online | phi-2 | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | cos_e_v1_11_generate_explanation_given_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | dbpedia_14_pick_one_category_for_the_following_text | phi-2 | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | super_glue_wsc_fixed_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | math_dataset_algebra__linear_1d_1_0_0 | phi-2 | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | duorc_ParaphraseRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | ropes_background_situation_middle | phi-2 | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | word_segment | phi-2 | sordonia/flan-10k-flat/word_segment | lora | | wmt16_translate_de_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | cnn_dailymail_3_4_0 | phi-2 | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | ropes_prompt_bottom_hint_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | wiqa_what_might_be_the_last_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | quail_no_prompt_id | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | cot_gsm8k | phi-2 | sordonia/flan-10k-flat/cot_gsm8k | lora | | duorc_ParaphraseRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quoref_What_Is_The_Answer | phi-2 | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | hellaswag_1_1_0 | phi-2 | sordonia/flan-10k-flat/hellaswag_1_1_0 | lora | | duorc_SelfRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | duorc_SelfRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | wiki_hop_original_explain_relation | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | dream_baseline | phi-2 | sordonia/flan-10k-flat/dream_baseline | lora | | definite_pronoun_resolution_1_1_0 | phi-2 | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | quail_context_description_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | quarel_do_not_use | phi-2 | sordonia/flan-10k-flat/quarel_do_not_use | lora | | quartz_paragraph_question_plain_concat | phi-2 | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | cot_ecqa | phi-2 | sordonia/flan-10k-flat/cot_ecqa | lora | | qasc_is_correct_2 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | ai2_arc_ARC_Challenge_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Challenge_1_0_0 | lora | | adversarial_qa_dbert_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | wiki_hop_original_choose_best_object_affirmative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | wiki_qa_Direct_Answer_to_Question | phi-2 | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | piqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/piqa_1_0_0 | lora | | sciq_Multiple_Choice_Question_First | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | phi-2 | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | duorc_ParaphraseRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | race_middle_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | drop_2_0_0 | phi-2 | sordonia/flan-10k-flat/drop_2_0_0 | lora | | kilt_tasks_hotpotqa_complex_question | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | natural_questions_open_1_0_0 | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | duorc_ParaphraseRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | cot_sensemaking_ii | phi-2 | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | kilt_tasks_hotpotqa_straighforward_qa | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | sciq_Direct_Question_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | adversarial_qa_dbidaf_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | race_high_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | cot_esnli | phi-2 | sordonia/flan-10k-flat/cot_esnli | lora | | cos_e_v1_11_question_option_description_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | quartz_use_info_from_paragraph_question | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | race_high_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | wiki_qa_Topic_Prediction_Answer_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | wmt14_translate_fr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | trivia_qa_rc_1_1_0 | phi-2 | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | kilt_tasks_hotpotqa_final_exam | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | ropes_prompt_mix | phi-2 | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | snli_1_1_0 | phi-2 | sordonia/flan-10k-flat/snli_1_1_0 | lora | | quarel_testing_students | phi-2 | sordonia/flan-10k-flat/quarel_testing_students | lora | | true_case | phi-2 | sordonia/flan-10k-flat/true_case | lora | | dream_generate_last_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | quoref_Context_Contains_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | gem_web_nlg_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | cot_strategyqa_ii | phi-2 | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | winogrande_1_1_0 | phi-2 | sordonia/flan-10k-flat/winogrande_1_1_0 | lora | | quartz_answer_question_based_on | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | quail_context_question_description_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | social_i_qa_Show_choices_and_generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | yelp_polarity_reviews_0_2_0 | phi-2 | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | adversarial_qa_dbidaf_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | phi-2 | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | duorc_ParaphraseRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | race_middle_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | qasc_qa_with_combined_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | duorc_ParaphraseRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | wiki_qa_exercise | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | dbpedia_14_given_a_choice_of_categories_ | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | quartz_given_the_fact_answer_the_q | phi-2 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | ropes_read_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | quail_context_question_description_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | race_middle_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | web_questions_short_general_knowledge_q | phi-2 | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | race_middle_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | wiki_bio_comprehension | phi-2 | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | wiki_hop_original_choose_best_object_affirmative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | kilt_tasks_hotpotqa_combining_facts | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | duorc_SelfRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | cos_e_v1_11_explain_why_human | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | quartz_use_info_from_question_paragraph | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | glue_sst2_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | race_high_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | duorc_SelfRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | anli_r2_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | web_questions_potential_correct_answer | phi-2 | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | quarel_heres_a_story | phi-2 | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | wiki_qa_Is_This_True_ | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | adversarial_qa_droberta_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | wiki_qa_Topic_Prediction_Question_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | race_high_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | quac_1_0_0 | phi-2 | sordonia/flan-10k-flat/quac_1_0_0 | lora | | quail_context_description_question_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | adversarial_qa_dbert_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | super_glue_copa_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | adversarial_qa_droberta_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | aeslc_1_0_0 | phi-2 | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | cot_strategyqa | phi-2 | sordonia/flan-10k-flat/cot_strategyqa | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | quail_description_context_question_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | adversarial_qa_dbert_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_bio_guess_person | phi-2 | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | app_reviews_convert_to_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiki_hop_original_generate_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | quail_context_question_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | sciq_Direct_Question | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | adversarial_qa_dbidaf_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | quoref_Read_And_Extract_ | phi-2 | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | race_middle_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | stream_aqua | phi-2 | sordonia/flan-10k-flat/stream_aqua | lora | | glue_qqp_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | web_questions_get_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | app_reviews_categorize_rating_using_review | phi-2 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | cot_esnli_ii | phi-2 | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quoref_Guess_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | quarel_choose_between | phi-2 | sordonia/flan-10k-flat/quarel_choose_between | lora | | gem_wiki_lingua_english_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | qasc_qa_with_separated_facts_5 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | ropes_plain_no_background | phi-2 | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | sciq_Multiple_Choice | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | fix_punct | phi-2 | sordonia/flan-10k-flat/fix_punct | lora | | cos_e_v1_11_question_description_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | quoref_Given_Context_Answer_Question | phi-2 | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | gem_common_gen_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_what_is_the_missing_first_step | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | cot_ecqa_ii | phi-2 | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | kilt_tasks_hotpotqa_formulate | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | qasc_is_correct_1 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | super_glue_cb_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | Last updated on: 2024-01-18 02:14:13+00:00
[ "SCIQ" ]
EleutherAI/pythia-1b-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-18T05:52:08Z"
2024-02-07T00:08:04+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-1b-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
EleutherAI/pythia-1.4b-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-18T05:53:04Z"
2024-02-07T00:08:16+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-1.4b-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
EleutherAI/pythia-2.8b-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-18T05:56:26Z"
2024-02-07T00:08:28+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-2.8b-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
taki0112/lora-trained-xl_craft-clay
taki0112
text-to-image
[ "diffusers", "tensorboard", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "lora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:openrail++", "region:us" ]
"2024-01-18T09:44:11Z"
2024-01-18T10:38:11+00:00
0
1
--- base_model: stabilityai/stable-diffusion-xl-base-1.0 license: openrail++ tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - text-to-image - diffusers - lora - template:sd-lora widget: - text: a deer in sks style output: url: image_0.png - text: a deer in sks style output: url: image_1.png - text: a deer in sks style output: url: image_2.png - text: a deer in sks style output: url: image_3.png instance_prompt: a dog in sks style --- # SDXL LoRA DreamBooth - taki0112/lora-trained-xl_craft-clay <Gallery /> ## Model description These are taki0112/lora-trained-xl_craft-clay LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. The weights were trained using [DreamBooth](https://dreambooth.github.io/). LoRA for the text encoder was enabled: False. Special VAE used for training: madebyollin/sdxl-vae-fp16-fix. ## Trigger words You should use a dog in sks style to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](taki0112/lora-trained-xl_craft-clay/tree/main) them in the Files & versions tab.
[ "CRAFT" ]
huizhang0110/hui-embedding
huizhang0110
null
[ "mteb", "model-index", "region:us" ]
"2024-01-18T10:24:23Z"
2024-11-26T05:15:59+00:00
0
0
--- tags: - mteb model-index: - name: no_model_name_available results: - task: type: STS dataset: name: MTEB STS22 (en) type: mteb/sts22-crosslingual-sts config: en split: test revision: de9d86b3b84231dc21f76c7b7af1f28e2f57f6e3 metrics: - type: cosine_pearson value: 66.2368177379181 - type: cosine_spearman value: 68.35446129213678 - type: euclidean_pearson value: 68.35832044207704 - type: euclidean_spearman value: 68.35446129213678 - type: main_score value: 68.35446129213678 - type: manhattan_pearson value: 68.70754373818515 - type: manhattan_spearman value: 68.2292889016414 - type: pearson value: 66.2368177379181 - type: spearman value: 68.35446129213678 - task: type: STS dataset: name: MTEB STS14 (default) type: mteb/sts14-sts config: default split: test revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375 metrics: - type: cosine_pearson value: 85.12461231748527 - type: cosine_spearman value: 83.78377223012504 - type: euclidean_pearson value: 84.84032421122767 - type: euclidean_spearman value: 83.78376987896931 - type: main_score value: 83.78377223012504 - type: manhattan_pearson value: 84.97174244411761 - type: manhattan_spearman value: 84.13202634643542 - type: pearson value: 85.12461231748527 - type: spearman value: 83.78377223012504 - task: type: Retrieval dataset: name: MTEB Touche2020 (default) type: mteb/touche2020 config: default split: test revision: a34f9a33db75fa0cbb21bb5cfc3dae8dc8bec93f metrics: - type: main_score value: 25.883 - type: map_at_1 value: 2.153 - type: map_at_10 value: 9.871 - type: map_at_100 value: 15.559000000000001 - type: map_at_1000 value: 17.183 - type: map_at_20 value: 12.552 - type: map_at_3 value: 5.493 - type: map_at_5 value: 7.85 - type: mrr_at_1 value: 30.612244897959183 - type: mrr_at_10 value: 48.89131843213475 - type: mrr_at_100 value: 49.6963561262702 - type: mrr_at_1000 value: 49.7010693279481 - type: mrr_at_20 value: 49.531452107982716 - type: mrr_at_3 value: 44.21768707482994 - type: mrr_at_5 value: 47.68707482993197 - type: nauc_map_at_1000_diff1 value: 25.31034571291797 - type: nauc_map_at_1000_max value: 34.51576312061718 - type: nauc_map_at_1000_std value: -4.906594382965329 - type: nauc_map_at_100_diff1 value: 25.837142212716476 - type: nauc_map_at_100_max value: 32.59407997636304 - type: nauc_map_at_100_std value: -10.217037670639481 - type: nauc_map_at_10_diff1 value: 33.21608048564407 - type: nauc_map_at_10_max value: 37.468380135605706 - type: nauc_map_at_10_std value: -20.46767738235632 - type: nauc_map_at_1_diff1 value: 32.281523854579106 - type: nauc_map_at_1_max value: 22.176737258675068 - type: nauc_map_at_1_std value: -25.07807730673564 - type: nauc_map_at_20_diff1 value: 30.866307166529584 - type: nauc_map_at_20_max value: 32.272418879076724 - type: nauc_map_at_20_std value: -20.40305363345012 - type: nauc_map_at_3_diff1 value: 30.88885591305534 - type: nauc_map_at_3_max value: 33.431908247176786 - type: nauc_map_at_3_std value: -19.503954175936993 - type: nauc_map_at_5_diff1 value: 34.08468180972433 - type: nauc_map_at_5_max value: 40.256459257111935 - type: nauc_map_at_5_std value: -18.56884658312989 - type: nauc_mrr_at_1000_diff1 value: 30.71882754790342 - type: nauc_mrr_at_1000_max value: 14.576101913381093 - type: nauc_mrr_at_1000_std value: -10.726757628242753 - type: nauc_mrr_at_100_diff1 value: 30.72979380373732 - type: nauc_mrr_at_100_max value: 14.58962334045265 - type: nauc_mrr_at_100_std value: -10.709231106839757 - type: nauc_mrr_at_10_diff1 value: 30.538894215258246 - type: nauc_mrr_at_10_max value: 13.803938970422532 - type: nauc_mrr_at_10_std value: -9.702168266086352 - type: nauc_mrr_at_1_diff1 value: 30.684478472773836 - type: nauc_mrr_at_1_max value: 17.71761545127753 - type: nauc_mrr_at_1_std value: -22.77705607353801 - type: nauc_mrr_at_20_diff1 value: 30.82506745472977 - type: nauc_mrr_at_20_max value: 14.664189943251788 - type: nauc_mrr_at_20_std value: -10.748922964408402 - type: nauc_mrr_at_3_diff1 value: 28.971974395355954 - type: nauc_mrr_at_3_max value: 14.14445297613165 - type: nauc_mrr_at_3_std value: -14.23741446560331 - type: nauc_mrr_at_5_diff1 value: 31.746911225636275 - type: nauc_mrr_at_5_max value: 14.268610321705955 - type: nauc_mrr_at_5_std value: -9.700708429060887 - type: nauc_ndcg_at_1000_diff1 value: 21.089489813171816 - type: nauc_ndcg_at_1000_max value: 28.175842354263764 - type: nauc_ndcg_at_1000_std value: 21.49424339507402 - type: nauc_ndcg_at_100_diff1 value: 19.8292750148825 - type: nauc_ndcg_at_100_max value: 17.123814348188652 - type: nauc_ndcg_at_100_std value: 6.051404399623092 - type: nauc_ndcg_at_10_diff1 value: 28.194702409547332 - type: nauc_ndcg_at_10_max value: 18.97062064198259 - type: nauc_ndcg_at_10_std value: -12.862439768903611 - type: nauc_ndcg_at_1_diff1 value: 30.684478472773836 - type: nauc_ndcg_at_1_max value: 17.71761545127753 - type: nauc_ndcg_at_1_std value: -22.77705607353801 - type: nauc_ndcg_at_20_diff1 value: 24.833493660655364 - type: nauc_ndcg_at_20_max value: 16.53068197823132 - type: nauc_ndcg_at_20_std value: -13.971353024276375 - type: nauc_ndcg_at_3_diff1 value: 29.840792656092052 - type: nauc_ndcg_at_3_max value: 18.823207152450045 - type: nauc_ndcg_at_3_std value: -12.753978007436833 - type: nauc_ndcg_at_5_diff1 value: 29.669577759746584 - type: nauc_ndcg_at_5_max value: 24.204580513440916 - type: nauc_ndcg_at_5_std value: -8.081655001819906 - type: nauc_precision_at_1000_diff1 value: -18.464873284114397 - type: nauc_precision_at_1000_max value: 21.495318097003405 - type: nauc_precision_at_1000_std value: 57.177192580535554 - type: nauc_precision_at_100_diff1 value: -4.067845048543001 - type: nauc_precision_at_100_max value: 13.157305810279249 - type: nauc_precision_at_100_std value: 51.20993669331124 - type: nauc_precision_at_10_diff1 value: 27.299848819776397 - type: nauc_precision_at_10_max value: 15.622698996242287 - type: nauc_precision_at_10_std value: -5.590347344162569 - type: nauc_precision_at_1_diff1 value: 30.684478472773836 - type: nauc_precision_at_1_max value: 17.71761545127753 - type: nauc_precision_at_1_std value: -22.77705607353801 - type: nauc_precision_at_20_diff1 value: 20.89429650870699 - type: nauc_precision_at_20_max value: 15.544972379682054 - type: nauc_precision_at_20_std value: 1.4293466620551607 - type: nauc_precision_at_3_diff1 value: 27.536001423592403 - type: nauc_precision_at_3_max value: 19.633139870619367 - type: nauc_precision_at_3_std value: -12.615890884253755 - type: nauc_precision_at_5_diff1 value: 27.120672981961334 - type: nauc_precision_at_5_max value: 27.279847435518494 - type: nauc_precision_at_5_std value: -4.87902522849883 - type: nauc_recall_at_1000_diff1 value: -2.8271060100732144 - type: nauc_recall_at_1000_max value: 20.480146626345764 - type: nauc_recall_at_1000_std value: 66.47919599815614 - type: nauc_recall_at_100_diff1 value: 12.101023414577305 - type: nauc_recall_at_100_max value: 10.468322459855992 - type: nauc_recall_at_100_std value: 18.442020075752115 - type: nauc_recall_at_10_diff1 value: 26.819559448061753 - type: nauc_recall_at_10_max value: 16.76558631205096 - type: nauc_recall_at_10_std value: -19.808438532692723 - type: nauc_recall_at_1_diff1 value: 32.281523854579106 - type: nauc_recall_at_1_max value: 22.176737258675068 - type: nauc_recall_at_1_std value: -25.07807730673564 - type: nauc_recall_at_20_diff1 value: 23.923306941170072 - type: nauc_recall_at_20_max value: 11.835367892374284 - type: nauc_recall_at_20_std value: -15.756707719745929 - type: nauc_recall_at_3_diff1 value: 24.836560375866597 - type: nauc_recall_at_3_max value: 23.378529137896617 - type: nauc_recall_at_3_std value: -20.181080283438245 - type: nauc_recall_at_5_diff1 value: 28.37846887432799 - type: nauc_recall_at_5_max value: 29.201940847759573 - type: nauc_recall_at_5_std value: -16.353497357324052 - type: ndcg_at_1 value: 30.612000000000002 - type: ndcg_at_10 value: 25.883 - type: ndcg_at_100 value: 36.213 - type: ndcg_at_1000 value: 47.952 - type: ndcg_at_20 value: 27.309 - type: ndcg_at_3 value: 30.532999999999998 - type: ndcg_at_5 value: 29.494999999999997 - type: precision_at_1 value: 30.612000000000002 - type: precision_at_10 value: 21.837 - type: precision_at_100 value: 7.286 - type: precision_at_1000 value: 1.488 - type: precision_at_20 value: 18.061 - type: precision_at_3 value: 31.293 - type: precision_at_5 value: 29.387999999999998 - type: recall_at_1 value: 2.153 - type: recall_at_10 value: 15.836 - type: recall_at_100 value: 44.199 - type: recall_at_1000 value: 79.809 - type: recall_at_20 value: 24.375 - type: recall_at_3 value: 6.729 - type: recall_at_5 value: 10.829 - task: type: Classification dataset: name: MTEB AmazonCounterfactualClassification (en) type: mteb/amazon_counterfactual config: en split: test revision: e8379541af4e31359cca9fbcf4b00f2671dba205 metrics: - type: accuracy value: 93.46268656716417 - type: ap value: 73.16905656173336 - type: ap_weighted value: 73.16905656173336 - type: f1 value: 89.9835927572066 - type: f1_weighted value: 93.578175628546 - type: main_score value: 93.46268656716417 - task: type: Classification dataset: name: MTEB AmazonPolarityClassification (default) type: mteb/amazon_polarity config: default split: test revision: e2d317d38cd51312af73b3d32a06d1a08b442046 metrics: - type: accuracy value: 96.64014999999999 - type: ap value: 94.75468802312224 - type: ap_weighted value: 94.75468802312224 - type: f1 value: 96.63929533118718 - type: f1_weighted value: 96.63929533118718 - type: main_score value: 96.64014999999999 - task: type: Classification dataset: name: MTEB AmazonReviewsClassification (en) type: mteb/amazon_reviews_multi config: en split: test revision: 1399c76144fd37290681b995c656ef9b2e06e26d metrics: - type: accuracy value: 63.970000000000006 - type: f1 value: 62.682765229278615 - type: f1_weighted value: 62.682765229278615 - type: main_score value: 63.970000000000006 - task: type: Retrieval dataset: name: MTEB ArguAna (default) type: mteb/arguana config: default split: test revision: c22ab2a51041ffd869aaddef7af8d8215647e41a metrics: - type: main_score value: 67.323 - type: map_at_1 value: 45.448 - type: map_at_10 value: 60.18599999999999 - type: map_at_100 value: 60.687999999999995 - type: map_at_1000 value: 60.690999999999995 - type: map_at_20 value: 60.563 - type: map_at_3 value: 57.053 - type: map_at_5 value: 58.867000000000004 - type: mrr_at_1 value: 45.59032716927454 - type: mrr_at_10 value: 60.22429384271503 - type: mrr_at_100 value: 60.72592259489321 - type: mrr_at_1000 value: 60.72916244163348 - type: mrr_at_20 value: 60.60091997479985 - type: mrr_at_3 value: 57.11237553342832 - type: mrr_at_5 value: 58.90469416785227 - type: nauc_map_at_1000_diff1 value: 33.068441925532255 - type: nauc_map_at_1000_max value: 10.276888386284378 - type: nauc_map_at_1000_std value: -16.78833416335307 - type: nauc_map_at_100_diff1 value: 33.07060210913634 - type: nauc_map_at_100_max value: 10.282642963249359 - type: nauc_map_at_100_std value: -16.781593405086635 - type: nauc_map_at_10_diff1 value: 32.83665966609534 - type: nauc_map_at_10_max value: 10.577416388110274 - type: nauc_map_at_10_std value: -16.505603731786895 - type: nauc_map_at_1_diff1 value: 38.109973823503346 - type: nauc_map_at_1_max value: 7.608684545790856 - type: nauc_map_at_1_std value: -17.893865428628576 - type: nauc_map_at_20_diff1 value: 33.044589968115 - type: nauc_map_at_20_max value: 10.375042647373576 - type: nauc_map_at_20_std value: -16.6822453639938 - type: nauc_map_at_3_diff1 value: 32.277891718391814 - type: nauc_map_at_3_max value: 9.850443641282443 - type: nauc_map_at_3_std value: -17.94797851381197 - type: nauc_map_at_5_diff1 value: 32.16092173570638 - type: nauc_map_at_5_max value: 10.209270409598851 - type: nauc_map_at_5_std value: -17.465881200007004 - type: nauc_mrr_at_1000_diff1 value: 32.827813536418006 - type: nauc_mrr_at_1000_max value: 10.087021629677352 - type: nauc_mrr_at_1000_std value: -16.967746341911923 - type: nauc_mrr_at_100_diff1 value: 32.83000077148736 - type: nauc_mrr_at_100_max value: 10.092796216302164 - type: nauc_mrr_at_100_std value: -16.960987105341093 - type: nauc_mrr_at_10_diff1 value: 32.60032888130517 - type: nauc_mrr_at_10_max value: 10.390784050073744 - type: nauc_mrr_at_10_std value: -16.681959182829477 - type: nauc_mrr_at_1_diff1 value: 37.728857246219 - type: nauc_mrr_at_1_max value: 7.4467908121287465 - type: nauc_mrr_at_1_std value: -18.30248538693518 - type: nauc_mrr_at_20_diff1 value: 32.80506350021981 - type: nauc_mrr_at_20_max value: 10.186006965165907 - type: nauc_mrr_at_20_std value: -16.86087660734542 - type: nauc_mrr_at_3_diff1 value: 32.19594731244019 - type: nauc_mrr_at_3_max value: 9.803200657757092 - type: nauc_mrr_at_3_std value: -18.11910256146044 - type: nauc_mrr_at_5_diff1 value: 31.933881085281225 - type: nauc_mrr_at_5_max value: 10.029923020334538 - type: nauc_mrr_at_5_std value: -17.635162099540345 - type: nauc_ndcg_at_1000_diff1 value: 32.518889050927 - type: nauc_ndcg_at_1000_max value: 10.875658070812662 - type: nauc_ndcg_at_1000_std value: -16.286324059189997 - type: nauc_ndcg_at_100_diff1 value: 32.57162076556983 - type: nauc_ndcg_at_100_max value: 11.011136236680544 - type: nauc_ndcg_at_100_std value: -16.1394614926114 - type: nauc_ndcg_at_10_diff1 value: 31.531506473288175 - type: nauc_ndcg_at_10_max value: 12.417307560165447 - type: nauc_ndcg_at_10_std value: -14.76971088523127 - type: nauc_ndcg_at_1_diff1 value: 38.109973823503346 - type: nauc_ndcg_at_1_max value: 7.608684545790856 - type: nauc_ndcg_at_1_std value: -17.893865428628576 - type: nauc_ndcg_at_20_diff1 value: 32.34260978744937 - type: nauc_ndcg_at_20_max value: 11.698122482769248 - type: nauc_ndcg_at_20_std value: -15.360551678773856 - type: nauc_ndcg_at_3_diff1 value: 30.412571299678465 - type: nauc_ndcg_at_3_max value: 10.694959789521832 - type: nauc_ndcg_at_3_std value: -18.138119030741954 - type: nauc_ndcg_at_5_diff1 value: 29.96746423000831 - type: nauc_ndcg_at_5_max value: 11.382928004181887 - type: nauc_ndcg_at_5_std value: -17.31473188362318 - type: nauc_precision_at_1000_diff1 value: 17.914806895369583 - type: nauc_precision_at_1000_max value: 24.542936056736938 - type: nauc_precision_at_1000_std value: 9.032925153517976 - type: nauc_precision_at_100_diff1 value: 36.40038451420755 - type: nauc_precision_at_100_max value: 44.12404870553998 - type: nauc_precision_at_100_std value: 23.899082906071847 - type: nauc_precision_at_10_diff1 value: 22.531117645662295 - type: nauc_precision_at_10_max value: 28.061598506640568 - type: nauc_precision_at_10_std value: 1.4390989358928021 - type: nauc_precision_at_1_diff1 value: 38.109973823503346 - type: nauc_precision_at_1_max value: 7.608684545790856 - type: nauc_precision_at_1_std value: -17.893865428628576 - type: nauc_precision_at_20_diff1 value: 27.52248228295167 - type: nauc_precision_at_20_max value: 31.544335924785592 - type: nauc_precision_at_20_std value: 7.8837210646197144 - type: nauc_precision_at_3_diff1 value: 23.746154368105525 - type: nauc_precision_at_3_max value: 13.770751722927347 - type: nauc_precision_at_3_std value: -18.895725316115847 - type: nauc_precision_at_5_diff1 value: 20.01291443486786 - type: nauc_precision_at_5_max value: 16.77718143645159 - type: nauc_precision_at_5_std value: -16.639028720975606 - type: nauc_recall_at_1000_diff1 value: 17.914806895363768 - type: nauc_recall_at_1000_max value: 24.542936056730795 - type: nauc_recall_at_1000_std value: 9.032925153519013 - type: nauc_recall_at_100_diff1 value: 36.400384514208625 - type: nauc_recall_at_100_max value: 44.1240487055395 - type: nauc_recall_at_100_std value: 23.899082906069786 - type: nauc_recall_at_10_diff1 value: 22.531117645662015 - type: nauc_recall_at_10_max value: 28.06159850664045 - type: nauc_recall_at_10_std value: 1.4390989358927389 - type: nauc_recall_at_1_diff1 value: 38.109973823503346 - type: nauc_recall_at_1_max value: 7.608684545790856 - type: nauc_recall_at_1_std value: -17.893865428628576 - type: nauc_recall_at_20_diff1 value: 27.52248228295106 - type: nauc_recall_at_20_max value: 31.54433592478498 - type: nauc_recall_at_20_std value: 7.883721064619468 - type: nauc_recall_at_3_diff1 value: 23.746154368105564 - type: nauc_recall_at_3_max value: 13.770751722927372 - type: nauc_recall_at_3_std value: -18.89572531611582 - type: nauc_recall_at_5_diff1 value: 20.01291443486774 - type: nauc_recall_at_5_max value: 16.77718143645164 - type: nauc_recall_at_5_std value: -16.639028720975606 - type: ndcg_at_1 value: 45.448 - type: ndcg_at_10 value: 67.323 - type: ndcg_at_100 value: 69.484 - type: ndcg_at_1000 value: 69.544 - type: ndcg_at_20 value: 68.644 - type: ndcg_at_3 value: 60.865 - type: ndcg_at_5 value: 64.125 - type: precision_at_1 value: 45.448 - type: precision_at_10 value: 8.969000000000001 - type: precision_at_100 value: 0.992 - type: precision_at_1000 value: 0.1 - type: precision_at_20 value: 4.74 - type: precision_at_3 value: 23.968999999999998 - type: precision_at_5 value: 15.959999999999999 - type: recall_at_1 value: 45.448 - type: recall_at_10 value: 89.687 - type: recall_at_100 value: 99.21799999999999 - type: recall_at_1000 value: 99.644 - type: recall_at_20 value: 94.808 - type: recall_at_3 value: 71.906 - type: recall_at_5 value: 79.801 - task: type: Clustering dataset: name: MTEB ArxivClusteringP2P (default) type: mteb/arxiv-clustering-p2p config: default split: test revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d metrics: - type: main_score value: 54.266038386390726 - type: v_measure value: 54.266038386390726 - type: v_measure_std value: 14.60711085325104 - task: type: Clustering dataset: name: MTEB ArxivClusteringS2S (default) type: mteb/arxiv-clustering-s2s config: default split: test revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53 metrics: - type: main_score value: 49.82450675178832 - type: v_measure value: 49.82450675178832 - type: v_measure_std value: 14.692705635234821 - task: type: Reranking dataset: name: MTEB AskUbuntuDupQuestions (default) type: mteb/askubuntudupquestions-reranking config: default split: test revision: 2000358ca161889fa9c082cb41daa8dcfb161a54 metrics: - type: main_score value: 63.50490353933854 - type: map value: 63.50490353933854 - type: mrr value: 76.79395858066218 - type: nAUC_map_diff1 value: 17.162853733308793 - type: nAUC_map_max value: 24.966054539639252 - type: nAUC_map_std value: 17.887481717389274 - type: nAUC_mrr_diff1 value: 19.033169471151794 - type: nAUC_mrr_max value: 37.52423297117104 - type: nAUC_mrr_std value: 13.636799292425081 - task: type: STS dataset: name: MTEB BIOSSES (default) type: mteb/biosses-sts config: default split: test revision: d3fb88f8f02e40887cd149695127462bbcf29b4a metrics: - type: cosine_pearson value: 82.07979992743572 - type: cosine_spearman value: 80.97112209037952 - type: euclidean_pearson value: 80.726157419205 - type: euclidean_spearman value: 80.97112209037952 - type: main_score value: 80.97112209037952 - type: manhattan_pearson value: 80.75553649447407 - type: manhattan_spearman value: 81.35366092726835 - type: pearson value: 82.07979992743572 - type: spearman value: 80.97112209037952 - task: type: Classification dataset: name: MTEB Banking77Classification (default) type: mteb/banking77 config: default split: test revision: 0fd18e25b25c072e09e0d92ab615fda904d66300 metrics: - type: accuracy value: 84.29870129870129 - type: f1 value: 83.68342680640103 - type: f1_weighted value: 83.68342680640104 - type: main_score value: 84.29870129870129 - task: type: Clustering dataset: name: MTEB BiorxivClusteringP2P (default) type: mteb/biorxiv-clustering-p2p config: default split: test revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40 metrics: - type: main_score value: 47.21939028186225 - type: v_measure value: 47.21939028186225 - type: v_measure_std value: 0.6399652619210745 - task: type: Clustering dataset: name: MTEB BiorxivClusteringS2S (default) type: mteb/biorxiv-clustering-s2s config: default split: test revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908 metrics: - type: main_score value: 43.904503083954765 - type: v_measure value: 43.904503083954765 - type: v_measure_std value: 0.6741506180366014 - task: type: Retrieval dataset: name: MTEB CQADupstackAndroidRetrieval (default) type: mteb/cqadupstack-android config: default split: test revision: f46a197baaae43b4f621051089b82a364682dfeb metrics: - type: main_score value: 50.441 - type: map_at_1 value: 30.581000000000003 - type: map_at_10 value: 43.536 - type: map_at_100 value: 45.086999999999996 - type: map_at_1000 value: 45.189 - type: map_at_20 value: 44.37 - type: map_at_3 value: 39.769 - type: map_at_5 value: 42.028999999999996 - type: mrr_at_1 value: 36.9098712446352 - type: mrr_at_10 value: 48.776483411676516 - type: mrr_at_100 value: 49.51782521965795 - type: mrr_at_1000 value: 49.5505304504549 - type: mrr_at_20 value: 49.20302191439193 - type: mrr_at_3 value: 45.85121602288984 - type: mrr_at_5 value: 47.696709585121575 - type: nauc_map_at_1000_diff1 value: 50.77442312287549 - type: nauc_map_at_1000_max value: 34.33647599905781 - type: nauc_map_at_1000_std value: -13.748516605067781 - type: nauc_map_at_100_diff1 value: 50.75894617435753 - type: nauc_map_at_100_max value: 34.35502812001472 - type: nauc_map_at_100_std value: -13.736841648468175 - type: nauc_map_at_10_diff1 value: 50.87929446622231 - type: nauc_map_at_10_max value: 34.27157508239978 - type: nauc_map_at_10_std value: -14.526407596674309 - type: nauc_map_at_1_diff1 value: 57.0909475560327 - type: nauc_map_at_1_max value: 32.2288149431883 - type: nauc_map_at_1_std value: -13.370874900310689 - type: nauc_map_at_20_diff1 value: 50.63798885082635 - type: nauc_map_at_20_max value: 34.26498561214445 - type: nauc_map_at_20_std value: -13.93188362561783 - type: nauc_map_at_3_diff1 value: 52.5737761085553 - type: nauc_map_at_3_max value: 33.76013333419806 - type: nauc_map_at_3_std value: -13.849008988263117 - type: nauc_map_at_5_diff1 value: 51.19968604216378 - type: nauc_map_at_5_max value: 33.54095507132505 - type: nauc_map_at_5_std value: -14.620211074645637 - type: nauc_mrr_at_1000_diff1 value: 48.38609356318301 - type: nauc_mrr_at_1000_max value: 33.98679377266471 - type: nauc_mrr_at_1000_std value: -13.759418374094038 - type: nauc_mrr_at_100_diff1 value: 48.37236116991555 - type: nauc_mrr_at_100_max value: 33.978575821483865 - type: nauc_mrr_at_100_std value: -13.748715391580502 - type: nauc_mrr_at_10_diff1 value: 48.20980705221954 - type: nauc_mrr_at_10_max value: 33.97030796624786 - type: nauc_mrr_at_10_std value: -14.023184119296047 - type: nauc_mrr_at_1_diff1 value: 52.835554088618565 - type: nauc_mrr_at_1_max value: 34.736747824514026 - type: nauc_mrr_at_1_std value: -14.782309133752246 - type: nauc_mrr_at_20_diff1 value: 48.185661393251586 - type: nauc_mrr_at_20_max value: 33.92181383095129 - type: nauc_mrr_at_20_std value: -13.749958473599438 - type: nauc_mrr_at_3_diff1 value: 49.06255086663413 - type: nauc_mrr_at_3_max value: 34.24245966485257 - type: nauc_mrr_at_3_std value: -14.121042079344855 - type: nauc_mrr_at_5_diff1 value: 48.02661914739764 - type: nauc_mrr_at_5_max value: 33.54319852163983 - type: nauc_mrr_at_5_std value: -14.40749724842102 - type: nauc_ndcg_at_1000_diff1 value: 48.93136634666757 - type: nauc_ndcg_at_1000_max value: 34.3178528230429 - type: nauc_ndcg_at_1000_std value: -11.95564837442876 - type: nauc_ndcg_at_100_diff1 value: 48.37542091427922 - type: nauc_ndcg_at_100_max value: 34.41374261950128 - type: nauc_ndcg_at_100_std value: -11.526876720456004 - type: nauc_ndcg_at_10_diff1 value: 48.27862794425633 - type: nauc_ndcg_at_10_max value: 34.06415523516767 - type: nauc_ndcg_at_10_std value: -14.602823441995778 - type: nauc_ndcg_at_1_diff1 value: 52.835554088618565 - type: nauc_ndcg_at_1_max value: 34.736747824514026 - type: nauc_ndcg_at_1_std value: -14.782309133752246 - type: nauc_ndcg_at_20_diff1 value: 47.6519433010848 - type: nauc_ndcg_at_20_max value: 33.800628012770034 - type: nauc_ndcg_at_20_std value: -12.852071902619322 - type: nauc_ndcg_at_3_diff1 value: 50.40632210084943 - type: nauc_ndcg_at_3_max value: 34.15616519598939 - type: nauc_ndcg_at_3_std value: -14.396914052394957 - type: nauc_ndcg_at_5_diff1 value: 48.2287768686924 - type: nauc_ndcg_at_5_max value: 32.68281782116356 - type: nauc_ndcg_at_5_std value: -15.15658424373146 - type: nauc_precision_at_1000_diff1 value: -16.822042402923493 - type: nauc_precision_at_1000_max value: -13.459387124925234 - type: nauc_precision_at_1000_std value: -4.684574162765856 - type: nauc_precision_at_100_diff1 value: -12.950405503358223 - type: nauc_precision_at_100_max value: -3.6973387744248694 - type: nauc_precision_at_100_std value: 1.0686120361051838 - type: nauc_precision_at_10_diff1 value: 5.680154771052575 - type: nauc_precision_at_10_max value: 17.15052960292624 - type: nauc_precision_at_10_std value: -8.839454848202234 - type: nauc_precision_at_1_diff1 value: 52.835554088618565 - type: nauc_precision_at_1_max value: 34.736747824514026 - type: nauc_precision_at_1_std value: -14.782309133752246 - type: nauc_precision_at_20_diff1 value: -4.147057156482801 - type: nauc_precision_at_20_max value: 8.943409955940282 - type: nauc_precision_at_20_std value: -1.1556219667822423 - type: nauc_precision_at_3_diff1 value: 28.31897232747757 - type: nauc_precision_at_3_max value: 28.82890100390525 - type: nauc_precision_at_3_std value: -14.111275428339304 - type: nauc_precision_at_5_diff1 value: 15.554085244274193 - type: nauc_precision_at_5_max value: 20.934501596265694 - type: nauc_precision_at_5_std value: -12.783947594197997 - type: nauc_recall_at_1000_diff1 value: 41.806713621678924 - type: nauc_recall_at_1000_max value: 45.83932148789026 - type: nauc_recall_at_1000_std value: 43.35832012725718 - type: nauc_recall_at_100_diff1 value: 33.259689472107425 - type: nauc_recall_at_100_max value: 34.680606932499764 - type: nauc_recall_at_100_std value: 13.022981792106265 - type: nauc_recall_at_10_diff1 value: 39.674972215899004 - type: nauc_recall_at_10_max value: 31.571411194709793 - type: nauc_recall_at_10_std value: -13.917597013140865 - type: nauc_recall_at_1_diff1 value: 57.0909475560327 - type: nauc_recall_at_1_max value: 32.2288149431883 - type: nauc_recall_at_1_std value: -13.370874900310689 - type: nauc_recall_at_20_diff1 value: 34.82477941545416 - type: nauc_recall_at_20_max value: 29.419097652367583 - type: nauc_recall_at_20_std value: -6.753466274035959 - type: nauc_recall_at_3_diff1 value: 47.05993622483161 - type: nauc_recall_at_3_max value: 31.788946521479673 - type: nauc_recall_at_3_std value: -12.589599804850593 - type: nauc_recall_at_5_diff1 value: 40.8980124793814 - type: nauc_recall_at_5_max value: 28.169478524380477 - type: nauc_recall_at_5_std value: -15.058399770454422 - type: ndcg_at_1 value: 36.91 - type: ndcg_at_10 value: 50.441 - type: ndcg_at_100 value: 55.986999999999995 - type: ndcg_at_1000 value: 57.50999999999999 - type: ndcg_at_20 value: 52.588 - type: ndcg_at_3 value: 45.039 - type: ndcg_at_5 value: 47.908 - type: precision_at_1 value: 36.91 - type: precision_at_10 value: 9.771 - type: precision_at_100 value: 1.5779999999999998 - type: precision_at_1000 value: 0.199 - type: precision_at_20 value: 5.808 - type: precision_at_3 value: 22.222 - type: precision_at_5 value: 16.28 - type: recall_at_1 value: 30.581000000000003 - type: recall_at_10 value: 64.43799999999999 - type: recall_at_100 value: 87.439 - type: recall_at_1000 value: 96.682 - type: recall_at_20 value: 72.021 - type: recall_at_3 value: 49.119 - type: recall_at_5 value: 56.650999999999996 - task: type: Retrieval dataset: name: MTEB CQADupstackEnglishRetrieval (default) type: mteb/cqadupstack-english config: default split: test revision: ad9991cb51e31e31e430383c75ffb2885547b5f0 metrics: - type: main_score value: 48.052 - type: map_at_1 value: 31.691000000000003 - type: map_at_10 value: 42.25 - type: map_at_100 value: 43.466 - type: map_at_1000 value: 43.592 - type: map_at_20 value: 42.925000000000004 - type: map_at_3 value: 39.196999999999996 - type: map_at_5 value: 40.837 - type: mrr_at_1 value: 40.0 - type: mrr_at_10 value: 48.40979172985544 - type: mrr_at_100 value: 49.01329345568664 - type: mrr_at_1000 value: 49.05317333733556 - type: mrr_at_20 value: 48.757963347938926 - type: mrr_at_3 value: 46.18895966029725 - type: mrr_at_5 value: 47.45647558386417 - type: nauc_map_at_1000_diff1 value: 52.63721197705168 - type: nauc_map_at_1000_max value: 34.927748424948255 - type: nauc_map_at_1000_std value: 1.0444719278570702 - type: nauc_map_at_100_diff1 value: 52.66002218018987 - type: nauc_map_at_100_max value: 34.89878215864321 - type: nauc_map_at_100_std value: 0.9008516460644733 - type: nauc_map_at_10_diff1 value: 52.889235851315775 - type: nauc_map_at_10_max value: 34.6922736480049 - type: nauc_map_at_10_std value: -0.6506284048085285 - type: nauc_map_at_1_diff1 value: 56.732243764713175 - type: nauc_map_at_1_max value: 30.49325212155099 - type: nauc_map_at_1_std value: -5.04800794470186 - type: nauc_map_at_20_diff1 value: 52.8862850560935 - type: nauc_map_at_20_max value: 34.829038965108325 - type: nauc_map_at_20_std value: 0.02178642495237562 - type: nauc_map_at_3_diff1 value: 53.608193764117 - type: nauc_map_at_3_max value: 33.53981267373349 - type: nauc_map_at_3_std value: -3.040418170003493 - type: nauc_map_at_5_diff1 value: 53.39851810143899 - type: nauc_map_at_5_max value: 34.5516659463275 - type: nauc_map_at_5_std value: -1.4969739346974889 - type: nauc_mrr_at_1000_diff1 value: 51.8960971254646 - type: nauc_mrr_at_1000_max value: 37.39091504745532 - type: nauc_mrr_at_1000_std value: 5.037970602087237 - type: nauc_mrr_at_100_diff1 value: 51.881385486300225 - type: nauc_mrr_at_100_max value: 37.38614133569158 - type: nauc_mrr_at_100_std value: 5.034384753845119 - type: nauc_mrr_at_10_diff1 value: 51.77335216991783 - type: nauc_mrr_at_10_max value: 37.61929128133669 - type: nauc_mrr_at_10_std value: 4.912421162621211 - type: nauc_mrr_at_1_diff1 value: 55.97789723641661 - type: nauc_mrr_at_1_max value: 38.07741378971052 - type: nauc_mrr_at_1_std value: 3.1114912067800407 - type: nauc_mrr_at_20_diff1 value: 51.924932204924964 - type: nauc_mrr_at_20_max value: 37.43188155675892 - type: nauc_mrr_at_20_std value: 4.912649497021889 - type: nauc_mrr_at_3_diff1 value: 52.62682614740191 - type: nauc_mrr_at_3_max value: 37.79696523235296 - type: nauc_mrr_at_3_std value: 4.297604310897065 - type: nauc_mrr_at_5_diff1 value: 51.93341098564305 - type: nauc_mrr_at_5_max value: 37.52261609729754 - type: nauc_mrr_at_5_std value: 4.798233142719436 - type: nauc_ndcg_at_1000_diff1 value: 50.48831175822571 - type: nauc_ndcg_at_1000_max value: 34.954324628161515 - type: nauc_ndcg_at_1000_std value: 5.914974932163024 - type: nauc_ndcg_at_100_diff1 value: 50.22642462713412 - type: nauc_ndcg_at_100_max value: 34.81144896724943 - type: nauc_ndcg_at_100_std value: 5.269669826884739 - type: nauc_ndcg_at_10_diff1 value: 50.638035087354346 - type: nauc_ndcg_at_10_max value: 35.548660617367744 - type: nauc_ndcg_at_10_std value: 2.757672387486977 - type: nauc_ndcg_at_1_diff1 value: 55.97789723641661 - type: nauc_ndcg_at_1_max value: 38.07741378971052 - type: nauc_ndcg_at_1_std value: 3.1114912067800407 - type: nauc_ndcg_at_20_diff1 value: 50.94165876070302 - type: nauc_ndcg_at_20_max value: 35.15720286509341 - type: nauc_ndcg_at_20_std value: 3.1700542955934177 - type: nauc_ndcg_at_3_diff1 value: 51.6668031483535 - type: nauc_ndcg_at_3_max value: 36.158392419704036 - type: nauc_ndcg_at_3_std value: 1.7945130542865129 - type: nauc_ndcg_at_5_diff1 value: 51.40374511387644 - type: nauc_ndcg_at_5_max value: 35.96747873017992 - type: nauc_ndcg_at_5_std value: 2.4750496896017036 - type: nauc_precision_at_1000_diff1 value: -14.459395103980057 - type: nauc_precision_at_1000_max value: 7.001254844374337 - type: nauc_precision_at_1000_std value: 38.87250799651196 - type: nauc_precision_at_100_diff1 value: -7.015008098259738 - type: nauc_precision_at_100_max value: 14.454169684224969 - type: nauc_precision_at_100_std value: 40.615163341328035 - type: nauc_precision_at_10_diff1 value: 14.105573590736311 - type: nauc_precision_at_10_max value: 27.637233565307927 - type: nauc_precision_at_10_std value: 24.80384513569725 - type: nauc_precision_at_1_diff1 value: 55.97789723641661 - type: nauc_precision_at_1_max value: 38.07741378971052 - type: nauc_precision_at_1_std value: 3.1114912067800407 - type: nauc_precision_at_20_diff1 value: 6.826222425028856 - type: nauc_precision_at_20_max value: 22.440750352931133 - type: nauc_precision_at_20_std value: 30.650961826400664 - type: nauc_precision_at_3_diff1 value: 33.56939227622927 - type: nauc_precision_at_3_max value: 35.81131949842977 - type: nauc_precision_at_3_std value: 13.39631093898116 - type: nauc_precision_at_5_diff1 value: 25.327171466051347 - type: nauc_precision_at_5_max value: 33.04313875843963 - type: nauc_precision_at_5_std value: 19.62165639744543 - type: nauc_recall_at_1000_diff1 value: 34.60133056300212 - type: nauc_recall_at_1000_max value: 21.161471663251515 - type: nauc_recall_at_1000_std value: 32.74321904619018 - type: nauc_recall_at_100_diff1 value: 36.43348185795896 - type: nauc_recall_at_100_max value: 25.704040738466205 - type: nauc_recall_at_100_std value: 17.990567238645156 - type: nauc_recall_at_10_diff1 value: 42.694617737297676 - type: nauc_recall_at_10_max value: 31.3298523819716 - type: nauc_recall_at_10_std value: 2.384843550540601 - type: nauc_recall_at_1_diff1 value: 56.732243764713175 - type: nauc_recall_at_1_max value: 30.49325212155099 - type: nauc_recall_at_1_std value: -5.04800794470186 - type: nauc_recall_at_20_diff1 value: 43.176907776217455 - type: nauc_recall_at_20_max value: 29.215827308916065 - type: nauc_recall_at_20_std value: 4.147830621064018 - type: nauc_recall_at_3_diff1 value: 48.35837999847456 - type: nauc_recall_at_3_max value: 31.92274839572281 - type: nauc_recall_at_3_std value: -2.714807149637697 - type: nauc_recall_at_5_diff1 value: 46.351251919981635 - type: nauc_recall_at_5_max value: 32.523267054288304 - type: nauc_recall_at_5_std value: 0.4952928034547165 - type: ndcg_at_1 value: 40.0 - type: ndcg_at_10 value: 48.052 - type: ndcg_at_100 value: 52.07000000000001 - type: ndcg_at_1000 value: 54.064 - type: ndcg_at_20 value: 49.626 - type: ndcg_at_3 value: 43.902 - type: ndcg_at_5 value: 45.701 - type: precision_at_1 value: 40.0 - type: precision_at_10 value: 9.203999999999999 - type: precision_at_100 value: 1.438 - type: precision_at_1000 value: 0.188 - type: precision_at_20 value: 5.376 - type: precision_at_3 value: 21.295 - type: precision_at_5 value: 15.082999999999998 - type: recall_at_1 value: 31.691000000000003 - type: recall_at_10 value: 57.859 - type: recall_at_100 value: 75.107 - type: recall_at_1000 value: 87.679 - type: recall_at_20 value: 63.698 - type: recall_at_3 value: 45.379000000000005 - type: recall_at_5 value: 50.556999999999995 - task: type: Retrieval dataset: name: MTEB CQADupstackGamingRetrieval (default) type: mteb/cqadupstack-gaming config: default split: test revision: 4885aa143210c98657558c04aaf3dc47cfb54340 metrics: - type: main_score value: 56.423 - type: map_at_1 value: 38.517 - type: map_at_10 value: 50.510999999999996 - type: map_at_100 value: 51.568000000000005 - type: map_at_1000 value: 51.625 - type: map_at_20 value: 51.157 - type: map_at_3 value: 46.861000000000004 - type: map_at_5 value: 49.138 - type: mrr_at_1 value: 44.32601880877743 - type: mrr_at_10 value: 54.006518385828805 - type: mrr_at_100 value: 54.71981356159521 - type: mrr_at_1000 value: 54.752151957526316 - type: mrr_at_20 value: 54.476952471748106 - type: mrr_at_3 value: 51.22257053291538 - type: mrr_at_5 value: 53.10658307210038 - type: nauc_map_at_1000_diff1 value: 46.99591363896863 - type: nauc_map_at_1000_max value: 36.24904681003381 - type: nauc_map_at_1000_std value: 5.266222409182784 - type: nauc_map_at_100_diff1 value: 46.98084517367969 - type: nauc_map_at_100_max value: 36.250423743143436 - type: nauc_map_at_100_std value: 5.28301960645914 - type: nauc_map_at_10_diff1 value: 47.06708588191781 - type: nauc_map_at_10_max value: 36.18369863811603 - type: nauc_map_at_10_std value: 4.540089547074973 - type: nauc_map_at_1_diff1 value: 50.48083670895566 - type: nauc_map_at_1_max value: 31.950879226720595 - type: nauc_map_at_1_std value: 0.8257060985358229 - type: nauc_map_at_20_diff1 value: 46.998991583288415 - type: nauc_map_at_20_max value: 36.22115301678039 - type: nauc_map_at_20_std value: 5.082303558564342 - type: nauc_map_at_3_diff1 value: 47.7005416643811 - type: nauc_map_at_3_max value: 35.88865564285155 - type: nauc_map_at_3_std value: 2.944332455222102 - type: nauc_map_at_5_diff1 value: 47.312929177575874 - type: nauc_map_at_5_max value: 35.862390825522844 - type: nauc_map_at_5_std value: 3.81274507266821 - type: nauc_mrr_at_1000_diff1 value: 46.6759837669438 - type: nauc_mrr_at_1000_max value: 36.70273979969576 - type: nauc_mrr_at_1000_std value: 5.372740994750759 - type: nauc_mrr_at_100_diff1 value: 46.675225471247536 - type: nauc_mrr_at_100_max value: 36.703302034269875 - type: nauc_mrr_at_100_std value: 5.389605566226372 - type: nauc_mrr_at_10_diff1 value: 46.50353044791382 - type: nauc_mrr_at_10_max value: 36.66777833991145 - type: nauc_mrr_at_10_std value: 5.243423563011071 - type: nauc_mrr_at_1_diff1 value: 49.02972042252377 - type: nauc_mrr_at_1_max value: 36.600499110729764 - type: nauc_mrr_at_1_std value: 2.5711258912407953 - type: nauc_mrr_at_20_diff1 value: 46.625296101632095 - type: nauc_mrr_at_20_max value: 36.678578716940855 - type: nauc_mrr_at_20_std value: 5.406361664314628 - type: nauc_mrr_at_3_diff1 value: 46.907538354326825 - type: nauc_mrr_at_3_max value: 36.91488611173621 - type: nauc_mrr_at_3_std value: 3.8761762810100473 - type: nauc_mrr_at_5_diff1 value: 46.774337072791255 - type: nauc_mrr_at_5_max value: 36.65454152790335 - type: nauc_mrr_at_5_std value: 4.753826902883721 - type: nauc_ndcg_at_1000_diff1 value: 46.312300114931396 - type: nauc_ndcg_at_1000_max value: 36.687577969558156 - type: nauc_ndcg_at_1000_std value: 8.04218255348285 - type: nauc_ndcg_at_100_diff1 value: 45.91371707529375 - type: nauc_ndcg_at_100_max value: 36.72698157851723 - type: nauc_ndcg_at_100_std value: 8.62715881456232 - type: nauc_ndcg_at_10_diff1 value: 45.70764954649013 - type: nauc_ndcg_at_10_max value: 36.42241644937269 - type: nauc_ndcg_at_10_std value: 6.793309697483774 - type: nauc_ndcg_at_1_diff1 value: 49.02972042252377 - type: nauc_ndcg_at_1_max value: 36.600499110729764 - type: nauc_ndcg_at_1_std value: 2.5711258912407953 - type: nauc_ndcg_at_20_diff1 value: 45.71253409870376 - type: nauc_ndcg_at_20_max value: 36.478750872235075 - type: nauc_ndcg_at_20_std value: 8.032852116533649 - type: nauc_ndcg_at_3_diff1 value: 46.5055405749989 - type: nauc_ndcg_at_3_max value: 36.55925519576953 - type: nauc_ndcg_at_3_std value: 4.01635426914171 - type: nauc_ndcg_at_5_diff1 value: 46.17076704583506 - type: nauc_ndcg_at_5_max value: 36.00194839608453 - type: nauc_ndcg_at_5_std value: 5.290651961116129 - type: nauc_precision_at_1000_diff1 value: -7.810936686028834 - type: nauc_precision_at_1000_max value: 2.4457731990668035 - type: nauc_precision_at_1000_std value: 15.244382957052343 - type: nauc_precision_at_100_diff1 value: -6.24711281837766 - type: nauc_precision_at_100_max value: 9.274662370763165 - type: nauc_precision_at_100_std value: 21.156495677287772 - type: nauc_precision_at_10_diff1 value: 11.673391020454202 - type: nauc_precision_at_10_max value: 23.642781032334476 - type: nauc_precision_at_10_std value: 15.428694149947766 - type: nauc_precision_at_1_diff1 value: 49.02972042252377 - type: nauc_precision_at_1_max value: 36.600499110729764 - type: nauc_precision_at_1_std value: 2.5711258912407953 - type: nauc_precision_at_20_diff1 value: 4.320523799516288 - type: nauc_precision_at_20_max value: 18.529188355144083 - type: nauc_precision_at_20_std value: 20.63811919289391 - type: nauc_precision_at_3_diff1 value: 28.81527179707099 - type: nauc_precision_at_3_max value: 34.12169505571048 - type: nauc_precision_at_3_std value: 8.264026657534398 - type: nauc_precision_at_5_diff1 value: 20.643744683841586 - type: nauc_precision_at_5_max value: 28.520212611799007 - type: nauc_precision_at_5_std value: 11.159926260802324 - type: nauc_recall_at_1000_diff1 value: 47.89843496456478 - type: nauc_recall_at_1000_max value: 48.19346950585018 - type: nauc_recall_at_1000_std value: 69.35955862460499 - type: nauc_recall_at_100_diff1 value: 38.5657115857761 - type: nauc_recall_at_100_max value: 39.1799100059013 - type: nauc_recall_at_100_std value: 37.26868224318161 - type: nauc_recall_at_10_diff1 value: 39.70450871697248 - type: nauc_recall_at_10_max value: 34.7230529664253 - type: nauc_recall_at_10_std value: 12.967503176766982 - type: nauc_recall_at_1_diff1 value: 50.48083670895566 - type: nauc_recall_at_1_max value: 31.950879226720595 - type: nauc_recall_at_1_std value: 0.8257060985358229 - type: nauc_recall_at_20_diff1 value: 38.52009076825669 - type: nauc_recall_at_20_max value: 35.067067464590004 - type: nauc_recall_at_20_std value: 21.157205479969708 - type: nauc_recall_at_3_diff1 value: 44.359044172441294 - type: nauc_recall_at_3_max value: 35.53948139234034 - type: nauc_recall_at_3_std value: 3.9964883607424118 - type: nauc_recall_at_5_diff1 value: 42.071462939937625 - type: nauc_recall_at_5_max value: 33.59544974420819 - type: nauc_recall_at_5_std value: 7.414365501450481 - type: ndcg_at_1 value: 44.326 - type: ndcg_at_10 value: 56.423 - type: ndcg_at_100 value: 60.626999999999995 - type: ndcg_at_1000 value: 61.78 - type: ndcg_at_20 value: 58.336 - type: ndcg_at_3 value: 50.32299999999999 - type: ndcg_at_5 value: 53.808 - type: precision_at_1 value: 44.326 - type: precision_at_10 value: 9.21 - type: precision_at_100 value: 1.2189999999999999 - type: precision_at_1000 value: 0.136 - type: precision_at_20 value: 5.176 - type: precision_at_3 value: 22.487 - type: precision_at_5 value: 15.9 - type: recall_at_1 value: 38.517 - type: recall_at_10 value: 70.291 - type: recall_at_100 value: 88.53999999999999 - type: recall_at_1000 value: 96.67 - type: recall_at_20 value: 77.459 - type: recall_at_3 value: 54.44 - type: recall_at_5 value: 62.863 - task: type: Retrieval dataset: name: MTEB CQADupstackGisRetrieval (default) type: mteb/cqadupstack-gis config: default split: test revision: 5003b3064772da1887988e05400cf3806fe491f2 metrics: - type: main_score value: 40.245999999999995 - type: map_at_1 value: 25.66 - type: map_at_10 value: 34.781 - type: map_at_100 value: 35.825 - type: map_at_1000 value: 35.900999999999996 - type: map_at_20 value: 35.312 - type: map_at_3 value: 31.649 - type: map_at_5 value: 33.446 - type: mrr_at_1 value: 27.683615819209038 - type: mrr_at_10 value: 36.88897856694467 - type: mrr_at_100 value: 37.7983608873635 - type: mrr_at_1000 value: 37.85659419024201 - type: mrr_at_20 value: 37.40279188480636 - type: mrr_at_3 value: 33.97363465160076 - type: mrr_at_5 value: 35.76459510357815 - type: nauc_map_at_1000_diff1 value: 39.85966393937221 - type: nauc_map_at_1000_max value: 27.627327546922082 - type: nauc_map_at_1000_std value: -2.437000048637541 - type: nauc_map_at_100_diff1 value: 39.84937090664403 - type: nauc_map_at_100_max value: 27.637564346944988 - type: nauc_map_at_100_std value: -2.4109683806023408 - type: nauc_map_at_10_diff1 value: 39.71424425034042 - type: nauc_map_at_10_max value: 27.872378136740437 - type: nauc_map_at_10_std value: -2.8569524387609566 - type: nauc_map_at_1_diff1 value: 45.91775607893774 - type: nauc_map_at_1_max value: 26.899324806364007 - type: nauc_map_at_1_std value: -5.498609993557515 - type: nauc_map_at_20_diff1 value: 39.883943198146106 - type: nauc_map_at_20_max value: 27.64309227085422 - type: nauc_map_at_20_std value: -2.5654741454169816 - type: nauc_map_at_3_diff1 value: 39.91753278618007 - type: nauc_map_at_3_max value: 27.11865653999877 - type: nauc_map_at_3_std value: -3.3286492180678384 - type: nauc_map_at_5_diff1 value: 39.6313699695734 - type: nauc_map_at_5_max value: 27.710946419917548 - type: nauc_map_at_5_std value: -2.920297786058066 - type: nauc_mrr_at_1000_diff1 value: 39.690653898179 - type: nauc_mrr_at_1000_max value: 27.18398591982711 - type: nauc_mrr_at_1000_std value: -2.606447174750376 - type: nauc_mrr_at_100_diff1 value: 39.689803477387656 - type: nauc_mrr_at_100_max value: 27.189479576677762 - type: nauc_mrr_at_100_std value: -2.570807442132712 - type: nauc_mrr_at_10_diff1 value: 39.399614568431915 - type: nauc_mrr_at_10_max value: 27.304654766506253 - type: nauc_mrr_at_10_std value: -2.8847962104122584 - type: nauc_mrr_at_1_diff1 value: 45.70161189197341 - type: nauc_mrr_at_1_max value: 27.02826003278829 - type: nauc_mrr_at_1_std value: -4.831200831009949 - type: nauc_mrr_at_20_diff1 value: 39.69394763509078 - type: nauc_mrr_at_20_max value: 27.201336203029232 - type: nauc_mrr_at_20_std value: -2.6871497640498765 - type: nauc_mrr_at_3_diff1 value: 39.220307350990346 - type: nauc_mrr_at_3_max value: 26.7053856409676 - type: nauc_mrr_at_3_std value: -3.2176631206275514 - type: nauc_mrr_at_5_diff1 value: 39.166108393948406 - type: nauc_mrr_at_5_max value: 27.084050550858557 - type: nauc_mrr_at_5_std value: -2.87556996749801 - type: nauc_ndcg_at_1000_diff1 value: 38.603857523266925 - type: nauc_ndcg_at_1000_max value: 27.45135486355824 - type: nauc_ndcg_at_1000_std value: -0.46660995944134603 - type: nauc_ndcg_at_100_diff1 value: 38.444207274649884 - type: nauc_ndcg_at_100_max value: 27.549884957721194 - type: nauc_ndcg_at_100_std value: 0.47388375830707924 - type: nauc_ndcg_at_10_diff1 value: 37.72567187058473 - type: nauc_ndcg_at_10_max value: 28.44081574137556 - type: nauc_ndcg_at_10_std value: -1.8534359145108148 - type: nauc_ndcg_at_1_diff1 value: 45.70161189197341 - type: nauc_ndcg_at_1_max value: 27.02826003278829 - type: nauc_ndcg_at_1_std value: -4.831200831009949 - type: nauc_ndcg_at_20_diff1 value: 38.44184854108953 - type: nauc_ndcg_at_20_max value: 27.679973388870614 - type: nauc_ndcg_at_20_std value: -0.898582155647988 - type: nauc_ndcg_at_3_diff1 value: 37.97088409897179 - type: nauc_ndcg_at_3_max value: 27.106412295185066 - type: nauc_ndcg_at_3_std value: -2.730164275362466 - type: nauc_ndcg_at_5_diff1 value: 37.37607068800825 - type: nauc_ndcg_at_5_max value: 27.9502784140078 - type: nauc_ndcg_at_5_std value: -2.0027830470055075 - type: nauc_precision_at_1000_diff1 value: 0.5286110453963512 - type: nauc_precision_at_1000_max value: -2.3318515785442813 - type: nauc_precision_at_1000_std value: 7.80079288314789 - type: nauc_precision_at_100_diff1 value: 13.667186642269913 - type: nauc_precision_at_100_max value: 9.942092016059734 - type: nauc_precision_at_100_std value: 12.50332782268112 - type: nauc_precision_at_10_diff1 value: 26.281496960169953 - type: nauc_precision_at_10_max value: 24.46085080936575 - type: nauc_precision_at_10_std value: 2.8074535999287322 - type: nauc_precision_at_1_diff1 value: 45.70161189197341 - type: nauc_precision_at_1_max value: 27.02826003278829 - type: nauc_precision_at_1_std value: -4.831200831009949 - type: nauc_precision_at_20_diff1 value: 25.585868175418412 - type: nauc_precision_at_20_max value: 19.640567118702023 - type: nauc_precision_at_20_std value: 7.0865072321039 - type: nauc_precision_at_3_diff1 value: 31.522547430107718 - type: nauc_precision_at_3_max value: 25.87424549883876 - type: nauc_precision_at_3_std value: -0.6508524960745287 - type: nauc_precision_at_5_diff1 value: 28.958347089826553 - type: nauc_precision_at_5_max value: 26.541109281414073 - type: nauc_precision_at_5_std value: 1.8354704960749444 - type: nauc_recall_at_1000_diff1 value: 25.74128427270277 - type: nauc_recall_at_1000_max value: 21.011729073123906 - type: nauc_recall_at_1000_std value: 29.766333163064136 - type: nauc_recall_at_100_diff1 value: 31.785700068938166 - type: nauc_recall_at_100_max value: 25.476332277500607 - type: nauc_recall_at_100_std value: 20.47758699126873 - type: nauc_recall_at_10_diff1 value: 31.186789594770264 - type: nauc_recall_at_10_max value: 30.23366916255125 - type: nauc_recall_at_10_std value: 1.0690146258142572 - type: nauc_recall_at_1_diff1 value: 45.91775607893774 - type: nauc_recall_at_1_max value: 26.899324806364007 - type: nauc_recall_at_1_std value: -5.498609993557515 - type: nauc_recall_at_20_diff1 value: 33.32210083840443 - type: nauc_recall_at_20_max value: 26.910239736720104 - type: nauc_recall_at_20_std value: 5.087368762147268 - type: nauc_recall_at_3_diff1 value: 32.3606502852846 - type: nauc_recall_at_3_max value: 26.86643484335275 - type: nauc_recall_at_3_std value: -0.9468851994313872 - type: nauc_recall_at_5_diff1 value: 30.58200958021165 - type: nauc_recall_at_5_max value: 28.81049824914163 - type: nauc_recall_at_5_std value: 0.40032324122162105 - type: ndcg_at_1 value: 27.683999999999997 - type: ndcg_at_10 value: 40.245999999999995 - type: ndcg_at_100 value: 45.506 - type: ndcg_at_1000 value: 47.461999999999996 - type: ndcg_at_20 value: 42.122 - type: ndcg_at_3 value: 34.209 - type: ndcg_at_5 value: 37.279 - type: precision_at_1 value: 27.683999999999997 - type: precision_at_10 value: 6.3839999999999995 - type: precision_at_100 value: 0.955 - type: precision_at_1000 value: 0.11499999999999999 - type: precision_at_20 value: 3.644 - type: precision_at_3 value: 14.539 - type: precision_at_5 value: 10.576 - type: recall_at_1 value: 25.66 - type: recall_at_10 value: 55.062999999999995 - type: recall_at_100 value: 79.38199999999999 - type: recall_at_1000 value: 94.233 - type: recall_at_20 value: 62.082 - type: recall_at_3 value: 39.078 - type: recall_at_5 value: 46.236 - task: type: Retrieval dataset: name: MTEB CQADupstackMathematicaRetrieval (default) type: mteb/cqadupstack-mathematica config: default split: test revision: 90fceea13679c63fe563ded68f3b6f06e50061de metrics: - type: main_score value: 33.599000000000004 - type: map_at_1 value: 17.586 - type: map_at_10 value: 27.32 - type: map_at_100 value: 28.799999999999997 - type: map_at_1000 value: 28.921000000000003 - type: map_at_20 value: 28.153 - type: map_at_3 value: 24.066000000000003 - type: map_at_5 value: 25.755 - type: mrr_at_1 value: 22.263681592039802 - type: mrr_at_10 value: 32.360469478006785 - type: mrr_at_100 value: 33.438437513063704 - type: mrr_at_1000 value: 33.497473884762094 - type: mrr_at_20 value: 33.04969965022066 - type: mrr_at_3 value: 29.415422885572156 - type: mrr_at_5 value: 30.99502487562189 - type: nauc_map_at_1000_diff1 value: 27.265553741965775 - type: nauc_map_at_1000_max value: 19.288555820766756 - type: nauc_map_at_1000_std value: 1.7933416168321978 - type: nauc_map_at_100_diff1 value: 27.260040534250695 - type: nauc_map_at_100_max value: 19.304892141398717 - type: nauc_map_at_100_std value: 1.8491919829209627 - type: nauc_map_at_10_diff1 value: 26.86944291051016 - type: nauc_map_at_10_max value: 18.92320522759212 - type: nauc_map_at_10_std value: 0.889749881009448 - type: nauc_map_at_1_diff1 value: 30.584243017075806 - type: nauc_map_at_1_max value: 13.491468441422066 - type: nauc_map_at_1_std value: -0.8751763698025199 - type: nauc_map_at_20_diff1 value: 27.227733914801732 - type: nauc_map_at_20_max value: 19.278767798642207 - type: nauc_map_at_20_std value: 1.4312898630264221 - type: nauc_map_at_3_diff1 value: 26.919576048874767 - type: nauc_map_at_3_max value: 18.312759768115967 - type: nauc_map_at_3_std value: -0.5642361688764358 - type: nauc_map_at_5_diff1 value: 27.04032364592226 - type: nauc_map_at_5_max value: 19.191923558129698 - type: nauc_map_at_5_std value: 0.14080066912052358 - type: nauc_mrr_at_1000_diff1 value: 27.136068664109448 - type: nauc_mrr_at_1000_max value: 22.022262336934386 - type: nauc_mrr_at_1000_std value: 3.3308260159907976 - type: nauc_mrr_at_100_diff1 value: 27.147288894737333 - type: nauc_mrr_at_100_max value: 22.02852436815082 - type: nauc_mrr_at_100_std value: 3.3550379360464526 - type: nauc_mrr_at_10_diff1 value: 26.79942635668937 - type: nauc_mrr_at_10_max value: 22.030637334814642 - type: nauc_mrr_at_10_std value: 2.867852159546408 - type: nauc_mrr_at_1_diff1 value: 29.595744930714023 - type: nauc_mrr_at_1_max value: 17.736581194275356 - type: nauc_mrr_at_1_std value: 0.2159541136892455 - type: nauc_mrr_at_20_diff1 value: 27.176010332894013 - type: nauc_mrr_at_20_max value: 22.13536761286141 - type: nauc_mrr_at_20_std value: 3.237439208098252 - type: nauc_mrr_at_3_diff1 value: 26.57000851252062 - type: nauc_mrr_at_3_max value: 21.747583860129698 - type: nauc_mrr_at_3_std value: 1.721057838979949 - type: nauc_mrr_at_5_diff1 value: 26.92551416387028 - type: nauc_mrr_at_5_max value: 22.42993672746205 - type: nauc_mrr_at_5_std value: 2.725843108347625 - type: nauc_ndcg_at_1000_diff1 value: 27.46739757065543 - type: nauc_ndcg_at_1000_max value: 21.041702596702677 - type: nauc_ndcg_at_1000_std value: 5.604780462883483 - type: nauc_ndcg_at_100_diff1 value: 27.652630070854155 - type: nauc_ndcg_at_100_max value: 21.81166185983459 - type: nauc_ndcg_at_100_std value: 6.698607031446962 - type: nauc_ndcg_at_10_diff1 value: 26.00697734505188 - type: nauc_ndcg_at_10_max value: 20.828161505269204 - type: nauc_ndcg_at_10_std value: 2.8399382855194033 - type: nauc_ndcg_at_1_diff1 value: 29.595744930714023 - type: nauc_ndcg_at_1_max value: 17.736581194275356 - type: nauc_ndcg_at_1_std value: 0.2159541136892455 - type: nauc_ndcg_at_20_diff1 value: 27.27378051779869 - type: nauc_ndcg_at_20_max value: 21.736204369394024 - type: nauc_ndcg_at_20_std value: 4.739094883714155 - type: nauc_ndcg_at_3_diff1 value: 26.57231894661191 - type: nauc_ndcg_at_3_max value: 20.93227880070676 - type: nauc_ndcg_at_3_std value: 0.024589831513874137 - type: nauc_ndcg_at_5_diff1 value: 26.600828085337064 - type: nauc_ndcg_at_5_max value: 21.773794661183416 - type: nauc_ndcg_at_5_std value: 1.5522574657313302 - type: nauc_precision_at_1000_diff1 value: 3.4210541212862537 - type: nauc_precision_at_1000_max value: 3.102103455114947 - type: nauc_precision_at_1000_std value: 1.7521716451583618 - type: nauc_precision_at_100_diff1 value: 11.443300353934575 - type: nauc_precision_at_100_max value: 14.660009751798997 - type: nauc_precision_at_100_std value: 12.668177644524992 - type: nauc_precision_at_10_diff1 value: 17.394001289019975 - type: nauc_precision_at_10_max value: 22.223278134383104 - type: nauc_precision_at_10_std value: 7.242926879010027 - type: nauc_precision_at_1_diff1 value: 29.595744930714023 - type: nauc_precision_at_1_max value: 17.736581194275356 - type: nauc_precision_at_1_std value: 0.2159541136892455 - type: nauc_precision_at_20_diff1 value: 17.43115026349507 - type: nauc_precision_at_20_max value: 21.47538261589186 - type: nauc_precision_at_20_std value: 10.237040595580279 - type: nauc_precision_at_3_diff1 value: 22.012366289647648 - type: nauc_precision_at_3_max value: 25.106312117807487 - type: nauc_precision_at_3_std value: 1.9995028727881818 - type: nauc_precision_at_5_diff1 value: 20.398546387324117 - type: nauc_precision_at_5_max value: 26.303228187054806 - type: nauc_precision_at_5_std value: 5.564748189759881 - type: nauc_recall_at_1000_diff1 value: 29.03481056576388 - type: nauc_recall_at_1000_max value: 17.81464147740126 - type: nauc_recall_at_1000_std value: 52.084053180233646 - type: nauc_recall_at_100_diff1 value: 28.23982991718224 - type: nauc_recall_at_100_max value: 26.168366200103815 - type: nauc_recall_at_100_std value: 28.36050476271469 - type: nauc_recall_at_10_diff1 value: 21.64818157792201 - type: nauc_recall_at_10_max value: 20.853972890132304 - type: nauc_recall_at_10_std value: 5.713144094583624 - type: nauc_recall_at_1_diff1 value: 30.584243017075806 - type: nauc_recall_at_1_max value: 13.491468441422066 - type: nauc_recall_at_1_std value: -0.8751763698025199 - type: nauc_recall_at_20_diff1 value: 25.370812868482425 - type: nauc_recall_at_20_max value: 23.485918438346335 - type: nauc_recall_at_20_std value: 13.06270351478354 - type: nauc_recall_at_3_diff1 value: 23.22354479137504 - type: nauc_recall_at_3_max value: 21.931741628585574 - type: nauc_recall_at_3_std value: 0.22215343527463874 - type: nauc_recall_at_5_diff1 value: 23.762779317387583 - type: nauc_recall_at_5_max value: 23.86601516024228 - type: nauc_recall_at_5_std value: 2.9938661959173722 - type: ndcg_at_1 value: 22.264 - type: ndcg_at_10 value: 33.599000000000004 - type: ndcg_at_100 value: 40.149 - type: ndcg_at_1000 value: 42.663000000000004 - type: ndcg_at_20 value: 36.329 - type: ndcg_at_3 value: 27.736 - type: ndcg_at_5 value: 30.219 - type: precision_at_1 value: 22.264 - type: precision_at_10 value: 6.542000000000001 - type: precision_at_100 value: 1.1320000000000001 - type: precision_at_1000 value: 0.147 - type: precision_at_20 value: 4.061 - type: precision_at_3 value: 14.013 - type: precision_at_5 value: 10.274 - type: recall_at_1 value: 17.586 - type: recall_at_10 value: 47.932 - type: recall_at_100 value: 75.958 - type: recall_at_1000 value: 93.512 - type: recall_at_20 value: 57.708999999999996 - type: recall_at_3 value: 31.46 - type: recall_at_5 value: 37.842 - task: type: Retrieval dataset: name: MTEB CQADupstackPhysicsRetrieval (default) type: mteb/cqadupstack-physics config: default split: test revision: 79531abbd1fb92d06c6d6315a0cbbbf5bb247ea4 metrics: - type: main_score value: 47.410000000000004 - type: map_at_1 value: 28.971999999999998 - type: map_at_10 value: 40.96 - type: map_at_100 value: 42.331 - type: map_at_1000 value: 42.441 - type: map_at_20 value: 41.742000000000004 - type: map_at_3 value: 37.393 - type: map_at_5 value: 39.407 - type: mrr_at_1 value: 36.092396535129936 - type: mrr_at_10 value: 46.3360297599951 - type: mrr_at_100 value: 47.12743931915083 - type: mrr_at_1000 value: 47.17149558717527 - type: mrr_at_20 value: 46.764900340591545 - type: mrr_at_3 value: 43.50336862367658 - type: mrr_at_5 value: 45.048123195380114 - type: nauc_map_at_1000_diff1 value: 48.875356927744015 - type: nauc_map_at_1000_max value: 26.846314785374048 - type: nauc_map_at_1000_std value: 0.23720537516106452 - type: nauc_map_at_100_diff1 value: 48.82990495193183 - type: nauc_map_at_100_max value: 26.843711103433947 - type: nauc_map_at_100_std value: 0.25686095628081784 - type: nauc_map_at_10_diff1 value: 48.72231653223161 - type: nauc_map_at_10_max value: 26.364353126291522 - type: nauc_map_at_10_std value: 0.01100727529750763 - type: nauc_map_at_1_diff1 value: 51.96344574112589 - type: nauc_map_at_1_max value: 27.671021546156044 - type: nauc_map_at_1_std value: -4.6808708326389805 - type: nauc_map_at_20_diff1 value: 48.870849394709346 - type: nauc_map_at_20_max value: 26.813670876224883 - type: nauc_map_at_20_std value: 0.2352693253381299 - type: nauc_map_at_3_diff1 value: 49.529072015100326 - type: nauc_map_at_3_max value: 27.77400144483059 - type: nauc_map_at_3_std value: -0.6453151987449416 - type: nauc_map_at_5_diff1 value: 49.20710807541119 - type: nauc_map_at_5_max value: 27.177488493074755 - type: nauc_map_at_5_std value: -0.25587902411032826 - type: nauc_mrr_at_1000_diff1 value: 48.498262710122425 - type: nauc_mrr_at_1000_max value: 26.11751051811526 - type: nauc_mrr_at_1000_std value: -0.7728285987105216 - type: nauc_mrr_at_100_diff1 value: 48.48746660434456 - type: nauc_mrr_at_100_max value: 26.115163451470647 - type: nauc_mrr_at_100_std value: -0.7480131276402198 - type: nauc_mrr_at_10_diff1 value: 48.43136858217138 - type: nauc_mrr_at_10_max value: 25.834024688307604 - type: nauc_mrr_at_10_std value: -0.9430552221216183 - type: nauc_mrr_at_1_diff1 value: 50.088598533173354 - type: nauc_mrr_at_1_max value: 27.648802533446197 - type: nauc_mrr_at_1_std value: -3.7628727544097984 - type: nauc_mrr_at_20_diff1 value: 48.473967578999215 - type: nauc_mrr_at_20_max value: 26.091998126081734 - type: nauc_mrr_at_20_std value: -0.7681300813435199 - type: nauc_mrr_at_3_diff1 value: 48.69610564249302 - type: nauc_mrr_at_3_max value: 27.373923497327624 - type: nauc_mrr_at_3_std value: -1.2747465922726908 - type: nauc_mrr_at_5_diff1 value: 48.53658899050662 - type: nauc_mrr_at_5_max value: 26.49833197267966 - type: nauc_mrr_at_5_std value: -0.8503446744063664 - type: nauc_ndcg_at_1000_diff1 value: 48.467870789955406 - type: nauc_ndcg_at_1000_max value: 26.04777255889547 - type: nauc_ndcg_at_1000_std value: 1.6645313343373058 - type: nauc_ndcg_at_100_diff1 value: 47.80533775872007 - type: nauc_ndcg_at_100_max value: 26.106122630999174 - type: nauc_ndcg_at_100_std value: 2.456751351490524 - type: nauc_ndcg_at_10_diff1 value: 47.57301034996511 - type: nauc_ndcg_at_10_max value: 24.379146216030552 - type: nauc_ndcg_at_10_std value: 1.2579497129670234 - type: nauc_ndcg_at_1_diff1 value: 50.088598533173354 - type: nauc_ndcg_at_1_max value: 27.648802533446197 - type: nauc_ndcg_at_1_std value: -3.7628727544097984 - type: nauc_ndcg_at_20_diff1 value: 47.87138595331042 - type: nauc_ndcg_at_20_max value: 25.648148427942452 - type: nauc_ndcg_at_20_std value: 2.1415614628731148 - type: nauc_ndcg_at_3_diff1 value: 48.40186907831459 - type: nauc_ndcg_at_3_max value: 27.015191238802633 - type: nauc_ndcg_at_3_std value: -0.28368565093265813 - type: nauc_ndcg_at_5_diff1 value: 48.43525178181797 - type: nauc_ndcg_at_5_max value: 26.033136810207125 - type: nauc_ndcg_at_5_std value: 0.5903319782637264 - type: nauc_precision_at_1000_diff1 value: -5.050204072247187 - type: nauc_precision_at_1000_max value: -1.706061543844424 - type: nauc_precision_at_1000_std value: 0.4935798158915392 - type: nauc_precision_at_100_diff1 value: 1.581628126436549 - type: nauc_precision_at_100_max value: 5.131864973231214 - type: nauc_precision_at_100_std value: 5.818785250601078 - type: nauc_precision_at_10_diff1 value: 17.826909304567316 - type: nauc_precision_at_10_max value: 10.047556755952215 - type: nauc_precision_at_10_std value: 5.828288769562702 - type: nauc_precision_at_1_diff1 value: 50.088598533173354 - type: nauc_precision_at_1_max value: 27.648802533446197 - type: nauc_precision_at_1_std value: -3.7628727544097984 - type: nauc_precision_at_20_diff1 value: 12.647456163352691 - type: nauc_precision_at_20_max value: 10.821622040896782 - type: nauc_precision_at_20_std value: 6.6782471423372405 - type: nauc_precision_at_3_diff1 value: 33.03366844205296 - type: nauc_precision_at_3_max value: 21.61654824915879 - type: nauc_precision_at_3_std value: 3.1117767791018403 - type: nauc_precision_at_5_diff1 value: 25.873738881952193 - type: nauc_precision_at_5_max value: 16.50897302333537 - type: nauc_precision_at_5_std value: 4.306391187216285 - type: nauc_recall_at_1000_diff1 value: 46.920916880807226 - type: nauc_recall_at_1000_max value: 18.93033931407027 - type: nauc_recall_at_1000_std value: 30.343625789039912 - type: nauc_recall_at_100_diff1 value: 36.99917690641126 - type: nauc_recall_at_100_max value: 21.9225154657857 - type: nauc_recall_at_100_std value: 20.18252525903621 - type: nauc_recall_at_10_diff1 value: 40.849017544403544 - type: nauc_recall_at_10_max value: 15.573050231627782 - type: nauc_recall_at_10_std value: 6.199240253446229 - type: nauc_recall_at_1_diff1 value: 51.96344574112589 - type: nauc_recall_at_1_max value: 27.671021546156044 - type: nauc_recall_at_1_std value: -4.6808708326389805 - type: nauc_recall_at_20_diff1 value: 41.15264820688897 - type: nauc_recall_at_20_max value: 19.50230922026062 - type: nauc_recall_at_20_std value: 11.139703256952268 - type: nauc_recall_at_3_diff1 value: 45.76731873825665 - type: nauc_recall_at_3_max value: 24.89502530374308 - type: nauc_recall_at_3_std value: 1.8833756018456458 - type: nauc_recall_at_5_diff1 value: 44.65491098304952 - type: nauc_recall_at_5_max value: 22.218813760031296 - type: nauc_recall_at_5_std value: 3.985541104014005 - type: ndcg_at_1 value: 36.092 - type: ndcg_at_10 value: 47.410000000000004 - type: ndcg_at_100 value: 52.829 - type: ndcg_at_1000 value: 54.736 - type: ndcg_at_20 value: 49.563 - type: ndcg_at_3 value: 41.724 - type: ndcg_at_5 value: 44.358 - type: precision_at_1 value: 36.092 - type: precision_at_10 value: 8.807 - type: precision_at_100 value: 1.336 - type: precision_at_1000 value: 0.166 - type: precision_at_20 value: 5.140000000000001 - type: precision_at_3 value: 20.244 - type: precision_at_5 value: 14.418000000000001 - type: recall_at_1 value: 28.971999999999998 - type: recall_at_10 value: 61.160000000000004 - type: recall_at_100 value: 83.60600000000001 - type: recall_at_1000 value: 95.696 - type: recall_at_20 value: 68.569 - type: recall_at_3 value: 45.269 - type: recall_at_5 value: 52.168000000000006 - task: type: Retrieval dataset: name: MTEB CQADupstackProgrammersRetrieval (default) type: mteb/cqadupstack-programmers config: default split: test revision: 6184bc1440d2dbc7612be22b50686b8826d22b32 metrics: - type: main_score value: 47.107 - type: map_at_1 value: 29.509999999999998 - type: map_at_10 value: 40.872 - type: map_at_100 value: 42.349 - type: map_at_1000 value: 42.441 - type: map_at_20 value: 41.743 - type: map_at_3 value: 37.174 - type: map_at_5 value: 39.232 - type: mrr_at_1 value: 36.41552511415525 - type: mrr_at_10 value: 46.10585634558234 - type: mrr_at_100 value: 47.04388507378313 - type: mrr_at_1000 value: 47.085192151800705 - type: mrr_at_20 value: 46.71053512338389 - type: mrr_at_3 value: 43.45509893455097 - type: mrr_at_5 value: 44.95624048706236 - type: nauc_map_at_1000_diff1 value: 47.17063593584487 - type: nauc_map_at_1000_max value: 35.6595416622568 - type: nauc_map_at_1000_std value: 1.882360177794315 - type: nauc_map_at_100_diff1 value: 47.141224266698956 - type: nauc_map_at_100_max value: 35.64890271359889 - type: nauc_map_at_100_std value: 1.909040104973397 - type: nauc_map_at_10_diff1 value: 47.0080172506447 - type: nauc_map_at_10_max value: 35.6271740598076 - type: nauc_map_at_10_std value: 1.5963064045936786 - type: nauc_map_at_1_diff1 value: 50.45344698710353 - type: nauc_map_at_1_max value: 34.1407536673108 - type: nauc_map_at_1_std value: -2.503693156800745 - type: nauc_map_at_20_diff1 value: 47.198265744193066 - type: nauc_map_at_20_max value: 35.59983959295096 - type: nauc_map_at_20_std value: 1.709629193868128 - type: nauc_map_at_3_diff1 value: 47.86035115628325 - type: nauc_map_at_3_max value: 33.91453079017758 - type: nauc_map_at_3_std value: -1.0125268264345189 - type: nauc_map_at_5_diff1 value: 47.57075430825601 - type: nauc_map_at_5_max value: 35.340050213538674 - type: nauc_map_at_5_std value: 0.565360701196888 - type: nauc_mrr_at_1000_diff1 value: 46.19502136847612 - type: nauc_mrr_at_1000_max value: 36.22787621665649 - type: nauc_mrr_at_1000_std value: 0.871072004307322 - type: nauc_mrr_at_100_diff1 value: 46.18202150096684 - type: nauc_mrr_at_100_max value: 36.2180237985802 - type: nauc_mrr_at_100_std value: 0.9124059695477915 - type: nauc_mrr_at_10_diff1 value: 46.016490051238904 - type: nauc_mrr_at_10_max value: 36.19342604363148 - type: nauc_mrr_at_10_std value: 0.9071792646788923 - type: nauc_mrr_at_1_diff1 value: 50.04822644213264 - type: nauc_mrr_at_1_max value: 38.40049220874411 - type: nauc_mrr_at_1_std value: -0.4331805170196953 - type: nauc_mrr_at_20_diff1 value: 46.154472362472056 - type: nauc_mrr_at_20_max value: 36.21027910317236 - type: nauc_mrr_at_20_std value: 0.7953830560986073 - type: nauc_mrr_at_3_diff1 value: 46.69193692769359 - type: nauc_mrr_at_3_max value: 36.09347122586123 - type: nauc_mrr_at_3_std value: -0.8314592280863028 - type: nauc_mrr_at_5_diff1 value: 46.36247573613005 - type: nauc_mrr_at_5_max value: 36.1332024555296 - type: nauc_mrr_at_5_std value: 0.08254138511110683 - type: nauc_ndcg_at_1000_diff1 value: 45.502836278293714 - type: nauc_ndcg_at_1000_max value: 35.46858202686828 - type: nauc_ndcg_at_1000_std value: 4.220566466316345 - type: nauc_ndcg_at_100_diff1 value: 44.97146510067551 - type: nauc_ndcg_at_100_max value: 35.20514680813267 - type: nauc_ndcg_at_100_std value: 5.3327590512159295 - type: nauc_ndcg_at_10_diff1 value: 44.77893725971796 - type: nauc_ndcg_at_10_max value: 35.30984188181181 - type: nauc_ndcg_at_10_std value: 3.643838626739208 - type: nauc_ndcg_at_1_diff1 value: 50.04822644213264 - type: nauc_ndcg_at_1_max value: 38.40049220874411 - type: nauc_ndcg_at_1_std value: -0.4331805170196953 - type: nauc_ndcg_at_20_diff1 value: 45.347579096264255 - type: nauc_ndcg_at_20_max value: 35.23900153649932 - type: nauc_ndcg_at_20_std value: 3.870932080127777 - type: nauc_ndcg_at_3_diff1 value: 45.73489028100815 - type: nauc_ndcg_at_3_max value: 33.456282441683534 - type: nauc_ndcg_at_3_std value: -0.4316489511717149 - type: nauc_ndcg_at_5_diff1 value: 45.64448042343172 - type: nauc_ndcg_at_5_max value: 34.82550522784654 - type: nauc_ndcg_at_5_std value: 1.625202909591719 - type: nauc_precision_at_1000_diff1 value: -11.082584414320458 - type: nauc_precision_at_1000_max value: -0.10525239966679063 - type: nauc_precision_at_1000_std value: 1.2049688164002124 - type: nauc_precision_at_100_diff1 value: -4.401663460913719 - type: nauc_precision_at_100_max value: 6.217580097767219 - type: nauc_precision_at_100_std value: 11.507170914733113 - type: nauc_precision_at_10_diff1 value: 15.316762589026817 - type: nauc_precision_at_10_max value: 24.094651080086884 - type: nauc_precision_at_10_std value: 14.997661405160551 - type: nauc_precision_at_1_diff1 value: 50.04822644213264 - type: nauc_precision_at_1_max value: 38.40049220874411 - type: nauc_precision_at_1_std value: -0.4331805170196953 - type: nauc_precision_at_20_diff1 value: 9.71755375786461 - type: nauc_precision_at_20_max value: 17.50245364945517 - type: nauc_precision_at_20_std value: 13.42442276093188 - type: nauc_precision_at_3_diff1 value: 33.92303910717078 - type: nauc_precision_at_3_max value: 31.577604822025844 - type: nauc_precision_at_3_std value: 4.225871813818534 - type: nauc_precision_at_5_diff1 value: 26.434077412071776 - type: nauc_precision_at_5_max value: 30.415493182198862 - type: nauc_precision_at_5_std value: 9.962587204978579 - type: nauc_recall_at_1000_diff1 value: 19.583141827294416 - type: nauc_recall_at_1000_max value: 25.331531875118163 - type: nauc_recall_at_1000_std value: 47.19745406634415 - type: nauc_recall_at_100_diff1 value: 28.38177952031043 - type: nauc_recall_at_100_max value: 27.04348472020136 - type: nauc_recall_at_100_std value: 32.64978369730068 - type: nauc_recall_at_10_diff1 value: 36.77645976843529 - type: nauc_recall_at_10_max value: 31.508362325677286 - type: nauc_recall_at_10_std value: 9.845183301924783 - type: nauc_recall_at_1_diff1 value: 50.45344698710353 - type: nauc_recall_at_1_max value: 34.1407536673108 - type: nauc_recall_at_1_std value: -2.503693156800745 - type: nauc_recall_at_20_diff1 value: 37.1245830532323 - type: nauc_recall_at_20_max value: 30.01404898730656 - type: nauc_recall_at_20_std value: 11.991031997571183 - type: nauc_recall_at_3_diff1 value: 41.50397374838714 - type: nauc_recall_at_3_max value: 28.605530200805894 - type: nauc_recall_at_3_std value: -0.2718652433235268 - type: nauc_recall_at_5_diff1 value: 39.85347018437693 - type: nauc_recall_at_5_max value: 30.8839592452558 - type: nauc_recall_at_5_std value: 4.6501737002456505 - type: ndcg_at_1 value: 36.416 - type: ndcg_at_10 value: 47.107 - type: ndcg_at_100 value: 52.998999999999995 - type: ndcg_at_1000 value: 54.647 - type: ndcg_at_20 value: 49.748 - type: ndcg_at_3 value: 41.555 - type: ndcg_at_5 value: 44.079 - type: precision_at_1 value: 36.416 - type: precision_at_10 value: 8.870000000000001 - type: precision_at_100 value: 1.381 - type: precision_at_1000 value: 0.168 - type: precision_at_20 value: 5.303 - type: precision_at_3 value: 19.901 - type: precision_at_5 value: 14.292 - type: recall_at_1 value: 29.509999999999998 - type: recall_at_10 value: 60.169 - type: recall_at_100 value: 84.745 - type: recall_at_1000 value: 95.515 - type: recall_at_20 value: 69.571 - type: recall_at_3 value: 44.751000000000005 - type: recall_at_5 value: 51.675000000000004 - task: type: Retrieval dataset: name: MTEB CQADupstackRetrieval (default) type: CQADupstackRetrieval_is_a_combined_dataset config: default split: test revision: CQADupstackRetrieval_is_a_combined_dataset metrics: - type: main_score value: 42.74816666666667 - type: ndcg_at_10 value: 42.74816666666667 - task: type: Retrieval dataset: name: MTEB CQADupstackStatsRetrieval (default) type: mteb/cqadupstack-stats config: default split: test revision: 65ac3a16b8e91f9cee4c9828cc7c335575432a2a metrics: - type: main_score value: 38.574999999999996 - type: map_at_1 value: 26.184 - type: map_at_10 value: 33.796 - type: map_at_100 value: 34.942 - type: map_at_1000 value: 35.027 - type: map_at_20 value: 34.400999999999996 - type: map_at_3 value: 31.247000000000003 - type: map_at_5 value: 32.618 - type: mrr_at_1 value: 29.447852760736197 - type: mrr_at_10 value: 36.7572913623527 - type: mrr_at_100 value: 37.75134813088362 - type: mrr_at_1000 value: 37.80936473092716 - type: mrr_at_20 value: 37.32794705157665 - type: mrr_at_3 value: 34.50920245398773 - type: mrr_at_5 value: 35.70552147239264 - type: nauc_map_at_1000_diff1 value: 41.03705943910151 - type: nauc_map_at_1000_max value: 31.212034574788223 - type: nauc_map_at_1000_std value: 5.0860482737872585 - type: nauc_map_at_100_diff1 value: 41.036679724384975 - type: nauc_map_at_100_max value: 31.22552462503921 - type: nauc_map_at_100_std value: 5.066442366854383 - type: nauc_map_at_10_diff1 value: 41.10544250629236 - type: nauc_map_at_10_max value: 31.36660549991616 - type: nauc_map_at_10_std value: 4.918474556305647 - type: nauc_map_at_1_diff1 value: 45.012348658917084 - type: nauc_map_at_1_max value: 32.97795481425923 - type: nauc_map_at_1_std value: 2.378119627191972 - type: nauc_map_at_20_diff1 value: 41.08979133865055 - type: nauc_map_at_20_max value: 31.215468276820857 - type: nauc_map_at_20_std value: 5.029740972247495 - type: nauc_map_at_3_diff1 value: 41.628234253590776 - type: nauc_map_at_3_max value: 32.53336359524941 - type: nauc_map_at_3_std value: 4.860348221405528 - type: nauc_map_at_5_diff1 value: 41.537709900129116 - type: nauc_map_at_5_max value: 32.276330681668654 - type: nauc_map_at_5_std value: 4.846181729651669 - type: nauc_mrr_at_1000_diff1 value: 42.29004874474518 - type: nauc_mrr_at_1000_max value: 31.307199153225735 - type: nauc_mrr_at_1000_std value: 4.605131934451417 - type: nauc_mrr_at_100_diff1 value: 42.280047109551546 - type: nauc_mrr_at_100_max value: 31.289947735731538 - type: nauc_mrr_at_100_std value: 4.582937582219149 - type: nauc_mrr_at_10_diff1 value: 42.34222112143596 - type: nauc_mrr_at_10_max value: 31.359940250531142 - type: nauc_mrr_at_10_std value: 4.453370071132275 - type: nauc_mrr_at_1_diff1 value: 45.95443881951325 - type: nauc_mrr_at_1_max value: 32.619135528025325 - type: nauc_mrr_at_1_std value: 2.052662449953393 - type: nauc_mrr_at_20_diff1 value: 42.26941002683479 - type: nauc_mrr_at_20_max value: 31.187438688521034 - type: nauc_mrr_at_20_std value: 4.5359475550655715 - type: nauc_mrr_at_3_diff1 value: 43.531839392022135 - type: nauc_mrr_at_3_max value: 32.21473960551518 - type: nauc_mrr_at_3_std value: 4.241677481952446 - type: nauc_mrr_at_5_diff1 value: 43.00448483997977 - type: nauc_mrr_at_5_max value: 31.936515068920237 - type: nauc_mrr_at_5_std value: 4.254613914320285 - type: nauc_ndcg_at_1000_diff1 value: 39.08960919974518 - type: nauc_ndcg_at_1000_max value: 30.08930269294802 - type: nauc_ndcg_at_1000_std value: 7.0902275178016225 - type: nauc_ndcg_at_100_diff1 value: 38.98713815279589 - type: nauc_ndcg_at_100_max value: 29.82144804645644 - type: nauc_ndcg_at_100_std value: 6.759601980797914 - type: nauc_ndcg_at_10_diff1 value: 39.418527591834795 - type: nauc_ndcg_at_10_max value: 30.08055189001222 - type: nauc_ndcg_at_10_std value: 5.721375611075414 - type: nauc_ndcg_at_1_diff1 value: 45.95443881951325 - type: nauc_ndcg_at_1_max value: 32.619135528025325 - type: nauc_ndcg_at_1_std value: 2.052662449953393 - type: nauc_ndcg_at_20_diff1 value: 39.05782103145853 - type: nauc_ndcg_at_20_max value: 29.49942876513546 - type: nauc_ndcg_at_20_std value: 6.34657136486055 - type: nauc_ndcg_at_3_diff1 value: 41.125063900984635 - type: nauc_ndcg_at_3_max value: 32.139095393552424 - type: nauc_ndcg_at_3_std value: 5.191262454292501 - type: nauc_ndcg_at_5_diff1 value: 40.717371213208544 - type: nauc_ndcg_at_5_max value: 31.774089542050117 - type: nauc_ndcg_at_5_std value: 5.223234037768828 - type: nauc_precision_at_1000_diff1 value: 0.006638310316025911 - type: nauc_precision_at_1000_max value: -9.546883023580094 - type: nauc_precision_at_1000_std value: -1.475622979214972 - type: nauc_precision_at_100_diff1 value: 11.010276773793507 - type: nauc_precision_at_100_max value: -0.08180253887926077 - type: nauc_precision_at_100_std value: 3.287046242664858 - type: nauc_precision_at_10_diff1 value: 27.262245018901698 - type: nauc_precision_at_10_max value: 16.2877591608577 - type: nauc_precision_at_10_std value: 5.839311010801853 - type: nauc_precision_at_1_diff1 value: 45.95443881951325 - type: nauc_precision_at_1_max value: 32.619135528025325 - type: nauc_precision_at_1_std value: 2.052662449953393 - type: nauc_precision_at_20_diff1 value: 22.408421524281493 - type: nauc_precision_at_20_max value: 10.077231751543565 - type: nauc_precision_at_20_std value: 6.236324897139737 - type: nauc_precision_at_3_diff1 value: 37.104186066630184 - type: nauc_precision_at_3_max value: 28.93970664421486 - type: nauc_precision_at_3_std value: 6.189175805816679 - type: nauc_precision_at_5_diff1 value: 33.481383755503344 - type: nauc_precision_at_5_max value: 24.574152076881976 - type: nauc_precision_at_5_std value: 5.787283838050964 - type: nauc_recall_at_1000_diff1 value: 13.749745478534466 - type: nauc_recall_at_1000_max value: 27.46595915304242 - type: nauc_recall_at_1000_std value: 43.337093159412746 - type: nauc_recall_at_100_diff1 value: 25.71608004026722 - type: nauc_recall_at_100_max value: 23.295361701635084 - type: nauc_recall_at_100_std value: 17.803464732957156 - type: nauc_recall_at_10_diff1 value: 31.44102657586473 - type: nauc_recall_at_10_max value: 25.636789857993808 - type: nauc_recall_at_10_std value: 8.690210156923568 - type: nauc_recall_at_1_diff1 value: 45.012348658917084 - type: nauc_recall_at_1_max value: 32.97795481425923 - type: nauc_recall_at_1_std value: 2.378119627191972 - type: nauc_recall_at_20_diff1 value: 29.75929214314049 - type: nauc_recall_at_20_max value: 22.919735188320487 - type: nauc_recall_at_20_std value: 11.567442926310765 - type: nauc_recall_at_3_diff1 value: 36.76334334420757 - type: nauc_recall_at_3_max value: 31.59129150974883 - type: nauc_recall_at_3_std value: 7.166175857606125 - type: nauc_recall_at_5_diff1 value: 35.13282132180025 - type: nauc_recall_at_5_max value: 30.350684835131553 - type: nauc_recall_at_5_std value: 7.142861662933231 - type: ndcg_at_1 value: 29.448 - type: ndcg_at_10 value: 38.574999999999996 - type: ndcg_at_100 value: 44.263999999999996 - type: ndcg_at_1000 value: 46.32 - type: ndcg_at_20 value: 40.628 - type: ndcg_at_3 value: 33.906 - type: ndcg_at_5 value: 36.03 - type: precision_at_1 value: 29.448 - type: precision_at_10 value: 6.166 - type: precision_at_100 value: 0.992 - type: precision_at_1000 value: 0.124 - type: precision_at_20 value: 3.627 - type: precision_at_3 value: 14.417 - type: precision_at_5 value: 10.184 - type: recall_at_1 value: 26.184 - type: recall_at_10 value: 50.339 - type: recall_at_100 value: 76.44300000000001 - type: recall_at_1000 value: 91.376 - type: recall_at_20 value: 57.94200000000001 - type: recall_at_3 value: 37.602000000000004 - type: recall_at_5 value: 42.708 - task: type: Retrieval dataset: name: MTEB CQADupstackTexRetrieval (default) type: mteb/cqadupstack-tex config: default split: test revision: 46989137a86843e03a6195de44b09deda022eec7 metrics: - type: main_score value: 30.446 - type: map_at_1 value: 16.888 - type: map_at_10 value: 25.169999999999998 - type: map_at_100 value: 26.432 - type: map_at_1000 value: 26.558 - type: map_at_20 value: 25.884 - type: map_at_3 value: 22.392 - type: map_at_5 value: 23.862 - type: mrr_at_1 value: 20.612525808671712 - type: mrr_at_10 value: 28.94379717934435 - type: mrr_at_100 value: 29.956026940991578 - type: mrr_at_1000 value: 30.030396196620284 - type: mrr_at_20 value: 29.543198182591162 - type: mrr_at_3 value: 26.34778618949303 - type: mrr_at_5 value: 27.725969259004412 - type: nauc_map_at_1000_diff1 value: 37.47997469039452 - type: nauc_map_at_1000_max value: 21.421931473251863 - type: nauc_map_at_1000_std value: 1.5690124401135066 - type: nauc_map_at_100_diff1 value: 37.478584664432304 - type: nauc_map_at_100_max value: 21.422143658021053 - type: nauc_map_at_100_std value: 1.566528536384738 - type: nauc_map_at_10_diff1 value: 37.63897487891542 - type: nauc_map_at_10_max value: 21.320030910527255 - type: nauc_map_at_10_std value: 1.0327570489355187 - type: nauc_map_at_1_diff1 value: 44.71163328349228 - type: nauc_map_at_1_max value: 22.446113464782112 - type: nauc_map_at_1_std value: -0.41970785990162957 - type: nauc_map_at_20_diff1 value: 37.60131283021686 - type: nauc_map_at_20_max value: 21.3691373960991 - type: nauc_map_at_20_std value: 1.2704576929639178 - type: nauc_map_at_3_diff1 value: 38.569300112130584 - type: nauc_map_at_3_max value: 21.599281592197645 - type: nauc_map_at_3_std value: 0.17312117243077374 - type: nauc_map_at_5_diff1 value: 38.003272593074534 - type: nauc_map_at_5_max value: 21.470587264514265 - type: nauc_map_at_5_std value: 0.8202467504176192 - type: nauc_mrr_at_1000_diff1 value: 36.40070606249303 - type: nauc_mrr_at_1000_max value: 20.918159385616235 - type: nauc_mrr_at_1000_std value: 1.4689044699534843 - type: nauc_mrr_at_100_diff1 value: 36.382723733435185 - type: nauc_mrr_at_100_max value: 20.914130048646378 - type: nauc_mrr_at_100_std value: 1.4695708792966349 - type: nauc_mrr_at_10_diff1 value: 36.39783865629839 - type: nauc_mrr_at_10_max value: 20.807844052080004 - type: nauc_mrr_at_10_std value: 1.0924977932781788 - type: nauc_mrr_at_1_diff1 value: 42.57454091873592 - type: nauc_mrr_at_1_max value: 21.672943617832036 - type: nauc_mrr_at_1_std value: -0.10189138615103883 - type: nauc_mrr_at_20_diff1 value: 36.3838114124106 - type: nauc_mrr_at_20_max value: 20.87264072376547 - type: nauc_mrr_at_20_std value: 1.3432553141494952 - type: nauc_mrr_at_3_diff1 value: 37.51571566935928 - type: nauc_mrr_at_3_max value: 21.19647468708375 - type: nauc_mrr_at_3_std value: 0.6277750127835567 - type: nauc_mrr_at_5_diff1 value: 36.87464282453542 - type: nauc_mrr_at_5_max value: 21.0704963624643 - type: nauc_mrr_at_5_std value: 0.9052912701483784 - type: nauc_ndcg_at_1000_diff1 value: 34.552555694361274 - type: nauc_ndcg_at_1000_max value: 21.259928579786788 - type: nauc_ndcg_at_1000_std value: 3.938486886570975 - type: nauc_ndcg_at_100_diff1 value: 34.37518593610454 - type: nauc_ndcg_at_100_max value: 21.182389588343348 - type: nauc_ndcg_at_100_std value: 4.3168049004409275 - type: nauc_ndcg_at_10_diff1 value: 35.211341808407504 - type: nauc_ndcg_at_10_max value: 20.84028975529198 - type: nauc_ndcg_at_10_std value: 1.8086338693039452 - type: nauc_ndcg_at_1_diff1 value: 42.57454091873592 - type: nauc_ndcg_at_1_max value: 21.672943617832036 - type: nauc_ndcg_at_1_std value: -0.10189138615103883 - type: nauc_ndcg_at_20_diff1 value: 35.00363891684754 - type: nauc_ndcg_at_20_max value: 20.922087179049363 - type: nauc_ndcg_at_20_std value: 2.660205273507509 - type: nauc_ndcg_at_3_diff1 value: 36.92485381743134 - type: nauc_ndcg_at_3_max value: 21.25737761098354 - type: nauc_ndcg_at_3_std value: 0.28798539980447146 - type: nauc_ndcg_at_5_diff1 value: 36.04502896798978 - type: nauc_ndcg_at_5_max value: 21.148648295149318 - type: nauc_ndcg_at_5_std value: 1.243003231031824 - type: nauc_precision_at_1000_diff1 value: -0.7759478803048101 - type: nauc_precision_at_1000_max value: 3.2826437330805502 - type: nauc_precision_at_1000_std value: 2.7787334076838173 - type: nauc_precision_at_100_diff1 value: 6.959433786637141 - type: nauc_precision_at_100_max value: 10.104545782506289 - type: nauc_precision_at_100_std value: 8.917540163713769 - type: nauc_precision_at_10_diff1 value: 22.003522151797437 - type: nauc_precision_at_10_max value: 16.164192732980553 - type: nauc_precision_at_10_std value: 3.275914834741683 - type: nauc_precision_at_1_diff1 value: 42.57454091873592 - type: nauc_precision_at_1_max value: 21.672943617832036 - type: nauc_precision_at_1_std value: -0.10189138615103883 - type: nauc_precision_at_20_diff1 value: 18.129059379732563 - type: nauc_precision_at_20_max value: 14.512665907788747 - type: nauc_precision_at_20_std value: 5.022877954638016 - type: nauc_precision_at_3_diff1 value: 29.98093015706584 - type: nauc_precision_at_3_max value: 19.728491902142636 - type: nauc_precision_at_3_std value: 1.4470534167918057 - type: nauc_precision_at_5_diff1 value: 26.50099880522309 - type: nauc_precision_at_5_max value: 18.138610189869738 - type: nauc_precision_at_5_std value: 2.551091667929808 - type: nauc_recall_at_1000_diff1 value: 16.96943824149726 - type: nauc_recall_at_1000_max value: 23.257191427293964 - type: nauc_recall_at_1000_std value: 24.9502432707826 - type: nauc_recall_at_100_diff1 value: 21.669754477643142 - type: nauc_recall_at_100_max value: 19.164964731074388 - type: nauc_recall_at_100_std value: 16.85249185076977 - type: nauc_recall_at_10_diff1 value: 27.551237362397828 - type: nauc_recall_at_10_max value: 18.28543172320463 - type: nauc_recall_at_10_std value: 3.5306584526336846 - type: nauc_recall_at_1_diff1 value: 44.71163328349228 - type: nauc_recall_at_1_max value: 22.446113464782112 - type: nauc_recall_at_1_std value: -0.41970785990162957 - type: nauc_recall_at_20_diff1 value: 26.271222471772326 - type: nauc_recall_at_20_max value: 18.12240775027493 - type: nauc_recall_at_20_std value: 6.607853337331698 - type: nauc_recall_at_3_diff1 value: 32.25185781878737 - type: nauc_recall_at_3_max value: 20.129371018198135 - type: nauc_recall_at_3_std value: 0.44779691255305437 - type: nauc_recall_at_5_diff1 value: 29.921019600841547 - type: nauc_recall_at_5_max value: 19.573769036363174 - type: nauc_recall_at_5_std value: 2.3711269481227277 - type: ndcg_at_1 value: 20.613 - type: ndcg_at_10 value: 30.446 - type: ndcg_at_100 value: 36.296 - type: ndcg_at_1000 value: 39.062999999999995 - type: ndcg_at_20 value: 32.756 - type: ndcg_at_3 value: 25.413000000000004 - type: ndcg_at_5 value: 27.61 - type: precision_at_1 value: 20.613 - type: precision_at_10 value: 5.785 - type: precision_at_100 value: 1.013 - type: precision_at_1000 value: 0.14400000000000002 - type: precision_at_20 value: 3.567 - type: precision_at_3 value: 12.216000000000001 - type: precision_at_5 value: 9.030000000000001 - type: recall_at_1 value: 16.888 - type: recall_at_10 value: 42.64 - type: recall_at_100 value: 68.771 - type: recall_at_1000 value: 88.018 - type: recall_at_20 value: 51.121 - type: recall_at_3 value: 28.505000000000003 - type: recall_at_5 value: 34.099000000000004 - task: type: Retrieval dataset: name: MTEB CQADupstackUnixRetrieval (default) type: mteb/cqadupstack-unix config: default split: test revision: 6c6430d3a6d36f8d2a829195bc5dc94d7e063e53 metrics: - type: main_score value: 42.283 - type: map_at_1 value: 26.326 - type: map_at_10 value: 36.515 - type: map_at_100 value: 37.832 - type: map_at_1000 value: 37.937 - type: map_at_20 value: 37.269999999999996 - type: map_at_3 value: 33.338 - type: map_at_5 value: 35.169 - type: mrr_at_1 value: 31.25 - type: mrr_at_10 value: 40.64062129827054 - type: mrr_at_100 value: 41.61370050636776 - type: mrr_at_1000 value: 41.67017916648837 - type: mrr_at_20 value: 41.18797257145757 - type: mrr_at_3 value: 37.93532338308455 - type: mrr_at_5 value: 39.55845771144274 - type: nauc_map_at_1000_diff1 value: 50.56683743231891 - type: nauc_map_at_1000_max value: 39.19969051920628 - type: nauc_map_at_1000_std value: -0.9554342712910933 - type: nauc_map_at_100_diff1 value: 50.544855645079345 - type: nauc_map_at_100_max value: 39.191632786859124 - type: nauc_map_at_100_std value: -0.962267377904227 - type: nauc_map_at_10_diff1 value: 50.74377893641161 - type: nauc_map_at_10_max value: 39.19344096202928 - type: nauc_map_at_10_std value: -1.22618961567281 - type: nauc_map_at_1_diff1 value: 57.492804899336434 - type: nauc_map_at_1_max value: 38.451829796562365 - type: nauc_map_at_1_std value: -2.2223809957991993 - type: nauc_map_at_20_diff1 value: 50.37379315470132 - type: nauc_map_at_20_max value: 39.15268041299702 - type: nauc_map_at_20_std value: -1.1542173582571251 - type: nauc_map_at_3_diff1 value: 52.114315032062265 - type: nauc_map_at_3_max value: 39.506520142355846 - type: nauc_map_at_3_std value: -1.136869114727129 - type: nauc_map_at_5_diff1 value: 51.137878020043615 - type: nauc_map_at_5_max value: 39.41597927774479 - type: nauc_map_at_5_std value: -1.414373986733375 - type: nauc_mrr_at_1000_diff1 value: 49.28345924937687 - type: nauc_mrr_at_1000_max value: 39.49024565022835 - type: nauc_mrr_at_1000_std value: -0.7389778084722739 - type: nauc_mrr_at_100_diff1 value: 49.25964062379304 - type: nauc_mrr_at_100_max value: 39.49625691927597 - type: nauc_mrr_at_100_std value: -0.7233812120562104 - type: nauc_mrr_at_10_diff1 value: 49.28005195010669 - type: nauc_mrr_at_10_max value: 39.502594291827194 - type: nauc_mrr_at_10_std value: -0.854578965146599 - type: nauc_mrr_at_1_diff1 value: 54.51968972219606 - type: nauc_mrr_at_1_max value: 38.985521654330725 - type: nauc_mrr_at_1_std value: -3.17796307755014 - type: nauc_mrr_at_20_diff1 value: 49.140932871712586 - type: nauc_mrr_at_20_max value: 39.44307540677674 - type: nauc_mrr_at_20_std value: -0.8396065147276742 - type: nauc_mrr_at_3_diff1 value: 50.04344397525612 - type: nauc_mrr_at_3_max value: 39.56654196970236 - type: nauc_mrr_at_3_std value: -1.2528287637913136 - type: nauc_mrr_at_5_diff1 value: 49.489373600446605 - type: nauc_mrr_at_5_max value: 39.659057230991316 - type: nauc_mrr_at_5_std value: -0.8720012571429344 - type: nauc_ndcg_at_1000_diff1 value: 48.748836050761405 - type: nauc_ndcg_at_1000_max value: 39.3457622357591 - type: nauc_ndcg_at_1000_std value: 1.1002389454170685 - type: nauc_ndcg_at_100_diff1 value: 48.22509167328338 - type: nauc_ndcg_at_100_max value: 39.3256932518086 - type: nauc_ndcg_at_100_std value: 1.438492059971218 - type: nauc_ndcg_at_10_diff1 value: 48.523357452437814 - type: nauc_ndcg_at_10_max value: 39.34471711241775 - type: nauc_ndcg_at_10_std value: -0.2137972110670513 - type: nauc_ndcg_at_1_diff1 value: 54.51968972219606 - type: nauc_ndcg_at_1_max value: 38.985521654330725 - type: nauc_ndcg_at_1_std value: -3.17796307755014 - type: nauc_ndcg_at_20_diff1 value: 47.51869995272205 - type: nauc_ndcg_at_20_max value: 39.30246710982855 - type: nauc_ndcg_at_20_std value: 0.1356281374446824 - type: nauc_ndcg_at_3_diff1 value: 50.12867016794126 - type: nauc_ndcg_at_3_max value: 39.4353732876648 - type: nauc_ndcg_at_3_std value: -0.9234551014485096 - type: nauc_ndcg_at_5_diff1 value: 49.10482448457108 - type: nauc_ndcg_at_5_max value: 39.604661308610275 - type: nauc_ndcg_at_5_std value: -0.7590788407730459 - type: nauc_precision_at_1000_diff1 value: -13.992133335670959 - type: nauc_precision_at_1000_max value: -7.214390627220537 - type: nauc_precision_at_1000_std value: 1.639261412748335 - type: nauc_precision_at_100_diff1 value: -0.557128351079009 - type: nauc_precision_at_100_max value: 7.486849612096312 - type: nauc_precision_at_100_std value: 7.1810501898680394 - type: nauc_precision_at_10_diff1 value: 21.213914544802844 - type: nauc_precision_at_10_max value: 25.864858450310546 - type: nauc_precision_at_10_std value: 0.39125389546740813 - type: nauc_precision_at_1_diff1 value: 54.51968972219606 - type: nauc_precision_at_1_max value: 38.985521654330725 - type: nauc_precision_at_1_std value: -3.17796307755014 - type: nauc_precision_at_20_diff1 value: 11.601304405847157 - type: nauc_precision_at_20_max value: 20.185407711622904 - type: nauc_precision_at_20_std value: 2.1916426458779488 - type: nauc_precision_at_3_diff1 value: 36.89740060012004 - type: nauc_precision_at_3_max value: 35.568914734056975 - type: nauc_precision_at_3_std value: 0.038850738796324405 - type: nauc_precision_at_5_diff1 value: 29.183999992678782 - type: nauc_precision_at_5_max value: 31.72969928353064 - type: nauc_precision_at_5_std value: -0.5629836594620032 - type: nauc_recall_at_1000_diff1 value: 37.261390414310384 - type: nauc_recall_at_1000_max value: 34.923735354550324 - type: nauc_recall_at_1000_std value: 45.97695232902403 - type: nauc_recall_at_100_diff1 value: 35.67925434563207 - type: nauc_recall_at_100_max value: 35.26178579038922 - type: nauc_recall_at_100_std value: 17.131274487036695 - type: nauc_recall_at_10_diff1 value: 40.90067655059736 - type: nauc_recall_at_10_max value: 36.79952710248241 - type: nauc_recall_at_10_std value: 2.716241775569224 - type: nauc_recall_at_1_diff1 value: 57.492804899336434 - type: nauc_recall_at_1_max value: 38.451829796562365 - type: nauc_recall_at_1_std value: -2.2223809957991993 - type: nauc_recall_at_20_diff1 value: 36.08583461458776 - type: nauc_recall_at_20_max value: 36.62990105037789 - type: nauc_recall_at_20_std value: 4.337305167037863 - type: nauc_recall_at_3_diff1 value: 46.41673012651659 - type: nauc_recall_at_3_max value: 38.842854844453505 - type: nauc_recall_at_3_std value: 0.8460605171745147 - type: nauc_recall_at_5_diff1 value: 43.29735456270288 - type: nauc_recall_at_5_max value: 38.51958912080913 - type: nauc_recall_at_5_std value: 1.1156101097663538 - type: ndcg_at_1 value: 31.25 - type: ndcg_at_10 value: 42.283 - type: ndcg_at_100 value: 48.067 - type: ndcg_at_1000 value: 50.246 - type: ndcg_at_20 value: 44.644 - type: ndcg_at_3 value: 36.858000000000004 - type: ndcg_at_5 value: 39.516 - type: precision_at_1 value: 31.25 - type: precision_at_10 value: 7.369000000000001 - type: precision_at_100 value: 1.137 - type: precision_at_1000 value: 0.14400000000000002 - type: precision_at_20 value: 4.328 - type: precision_at_3 value: 17.071 - type: precision_at_5 value: 12.257 - type: recall_at_1 value: 26.326 - type: recall_at_10 value: 55.689 - type: recall_at_100 value: 80.60000000000001 - type: recall_at_1000 value: 95.33500000000001 - type: recall_at_20 value: 64.229 - type: recall_at_3 value: 40.836 - type: recall_at_5 value: 47.577000000000005 - task: type: Retrieval dataset: name: MTEB CQADupstackWebmastersRetrieval (default) type: mteb/cqadupstack-webmasters config: default split: test revision: 160c094312a0e1facb97e55eeddb698c0abe3571 metrics: - type: main_score value: 42.841 - type: map_at_1 value: 26.180999999999997 - type: map_at_10 value: 36.370000000000005 - type: map_at_100 value: 38.143 - type: map_at_1000 value: 38.344 - type: map_at_20 value: 37.333 - type: map_at_3 value: 33.061 - type: map_at_5 value: 34.776 - type: mrr_at_1 value: 31.422924901185773 - type: mrr_at_10 value: 40.798513081121804 - type: mrr_at_100 value: 41.768341726195324 - type: mrr_at_1000 value: 41.814384046688836 - type: mrr_at_20 value: 41.39358410960642 - type: mrr_at_3 value: 37.91172595520424 - type: mrr_at_5 value: 39.39393939393942 - type: nauc_map_at_1000_diff1 value: 54.03266808787207 - type: nauc_map_at_1000_max value: 27.336058137590257 - type: nauc_map_at_1000_std value: 7.929483408524749 - type: nauc_map_at_100_diff1 value: 54.07605617892997 - type: nauc_map_at_100_max value: 27.420933517205988 - type: nauc_map_at_100_std value: 7.654364598186232 - type: nauc_map_at_10_diff1 value: 54.49638487392654 - type: nauc_map_at_10_max value: 26.95803941812555 - type: nauc_map_at_10_std value: 6.0656481626678955 - type: nauc_map_at_1_diff1 value: 60.62163331093275 - type: nauc_map_at_1_max value: 30.354161137182604 - type: nauc_map_at_1_std value: 3.283563596176243 - type: nauc_map_at_20_diff1 value: 54.2171414323596 - type: nauc_map_at_20_max value: 27.284531333468713 - type: nauc_map_at_20_std value: 6.98275284446578 - type: nauc_map_at_3_diff1 value: 55.48999072237882 - type: nauc_map_at_3_max value: 27.87434380647368 - type: nauc_map_at_3_std value: 5.868275382905556 - type: nauc_map_at_5_diff1 value: 54.84718663927504 - type: nauc_map_at_5_max value: 26.76192258450303 - type: nauc_map_at_5_std value: 4.739255945404961 - type: nauc_mrr_at_1000_diff1 value: 53.90866989000705 - type: nauc_mrr_at_1000_max value: 28.600059918390247 - type: nauc_mrr_at_1000_std value: 9.096507718338657 - type: nauc_mrr_at_100_diff1 value: 53.902988075226396 - type: nauc_mrr_at_100_max value: 28.599830953942174 - type: nauc_mrr_at_100_std value: 9.106284426792636 - type: nauc_mrr_at_10_diff1 value: 53.80346272826417 - type: nauc_mrr_at_10_max value: 28.281963295521706 - type: nauc_mrr_at_10_std value: 8.759210459852863 - type: nauc_mrr_at_1_diff1 value: 60.080144505628354 - type: nauc_mrr_at_1_max value: 33.74016395865226 - type: nauc_mrr_at_1_std value: 7.6714142708021305 - type: nauc_mrr_at_20_diff1 value: 53.760177497884406 - type: nauc_mrr_at_20_max value: 28.463215939799813 - type: nauc_mrr_at_20_std value: 9.068314971833093 - type: nauc_mrr_at_3_diff1 value: 54.41179314982579 - type: nauc_mrr_at_3_max value: 29.01231966941189 - type: nauc_mrr_at_3_std value: 9.383760609453352 - type: nauc_mrr_at_5_diff1 value: 54.261154767714515 - type: nauc_mrr_at_5_max value: 28.187796326709314 - type: nauc_mrr_at_5_std value: 8.324984381963386 - type: nauc_ndcg_at_1000_diff1 value: 52.16756830119805 - type: nauc_ndcg_at_1000_max value: 27.47333072396369 - type: nauc_ndcg_at_1000_std value: 10.433977027658207 - type: nauc_ndcg_at_100_diff1 value: 51.67893475997602 - type: nauc_ndcg_at_100_max value: 27.364432612842776 - type: nauc_ndcg_at_100_std value: 10.418878470827911 - type: nauc_ndcg_at_10_diff1 value: 51.455066768364546 - type: nauc_ndcg_at_10_max value: 24.86204769904609 - type: nauc_ndcg_at_10_std value: 7.975685972633213 - type: nauc_ndcg_at_1_diff1 value: 60.080144505628354 - type: nauc_ndcg_at_1_max value: 33.74016395865226 - type: nauc_ndcg_at_1_std value: 7.6714142708021305 - type: nauc_ndcg_at_20_diff1 value: 51.135229230296154 - type: nauc_ndcg_at_20_max value: 25.718284057364894 - type: nauc_ndcg_at_20_std value: 9.289363271312794 - type: nauc_ndcg_at_3_diff1 value: 52.70782059846899 - type: nauc_ndcg_at_3_max value: 27.470104306225863 - type: nauc_ndcg_at_3_std value: 8.98582220953654 - type: nauc_ndcg_at_5_diff1 value: 52.13622381467935 - type: nauc_ndcg_at_5_max value: 25.012072634464516 - type: nauc_ndcg_at_5_std value: 6.400559275913626 - type: nauc_precision_at_1000_diff1 value: -8.068455064670975 - type: nauc_precision_at_1000_max value: -10.387599717496192 - type: nauc_precision_at_1000_std value: 29.28771717137362 - type: nauc_precision_at_100_diff1 value: -4.542486688876828 - type: nauc_precision_at_100_max value: 2.2213727010948805 - type: nauc_precision_at_100_std value: 28.27046916836265 - type: nauc_precision_at_10_diff1 value: 19.415176505821286 - type: nauc_precision_at_10_max value: 13.444503991503346 - type: nauc_precision_at_10_std value: 16.810075843089322 - type: nauc_precision_at_1_diff1 value: 60.080144505628354 - type: nauc_precision_at_1_max value: 33.74016395865226 - type: nauc_precision_at_1_std value: 7.6714142708021305 - type: nauc_precision_at_20_diff1 value: 7.891942509311732 - type: nauc_precision_at_20_max value: 9.684197810455526 - type: nauc_precision_at_20_std value: 22.88953757757932 - type: nauc_precision_at_3_diff1 value: 37.07359628126754 - type: nauc_precision_at_3_max value: 23.182518856006016 - type: nauc_precision_at_3_std value: 15.043709459451618 - type: nauc_precision_at_5_diff1 value: 30.603525439923317 - type: nauc_precision_at_5_max value: 17.887460487183446 - type: nauc_precision_at_5_std value: 10.354003595459048 - type: nauc_recall_at_1000_diff1 value: 37.24937924148794 - type: nauc_recall_at_1000_max value: 27.116312668851744 - type: nauc_recall_at_1000_std value: 53.85172781866263 - type: nauc_recall_at_100_diff1 value: 36.95341517350607 - type: nauc_recall_at_100_max value: 26.388323872148362 - type: nauc_recall_at_100_std value: 25.552378739251036 - type: nauc_recall_at_10_diff1 value: 40.71842158421213 - type: nauc_recall_at_10_max value: 16.378208729794906 - type: nauc_recall_at_10_std value: 7.038163226525162 - type: nauc_recall_at_1_diff1 value: 60.62163331093275 - type: nauc_recall_at_1_max value: 30.354161137182604 - type: nauc_recall_at_1_std value: 3.283563596176243 - type: nauc_recall_at_20_diff1 value: 37.67229343743934 - type: nauc_recall_at_20_max value: 19.09861858622759 - type: nauc_recall_at_20_std value: 12.498129510164299 - type: nauc_recall_at_3_diff1 value: 47.38926382155088 - type: nauc_recall_at_3_max value: 21.835926284104218 - type: nauc_recall_at_3_std value: 6.956536082796651 - type: nauc_recall_at_5_diff1 value: 44.52691027171522 - type: nauc_recall_at_5_max value: 16.60678467044489 - type: nauc_recall_at_5_std value: 2.751824192702687 - type: ndcg_at_1 value: 31.423000000000002 - type: ndcg_at_10 value: 42.841 - type: ndcg_at_100 value: 49.003 - type: ndcg_at_1000 value: 51.117999999999995 - type: ndcg_at_20 value: 45.273 - type: ndcg_at_3 value: 37.469 - type: ndcg_at_5 value: 39.841 - type: precision_at_1 value: 31.423000000000002 - type: precision_at_10 value: 8.419 - type: precision_at_100 value: 1.638 - type: precision_at_1000 value: 0.243 - type: precision_at_20 value: 5.375 - type: precision_at_3 value: 17.852 - type: precision_at_5 value: 12.964 - type: recall_at_1 value: 26.180999999999997 - type: recall_at_10 value: 55.564 - type: recall_at_100 value: 83.22500000000001 - type: recall_at_1000 value: 96.124 - type: recall_at_20 value: 64.68199999999999 - type: recall_at_3 value: 40.28 - type: recall_at_5 value: 46.535 - task: type: Retrieval dataset: name: MTEB CQADupstackWordpressRetrieval (default) type: mteb/cqadupstack-wordpress config: default split: test revision: 4ffe81d471b1924886b33c7567bfb200e9eec5c4 metrics: - type: main_score value: 35.555 - type: map_at_1 value: 22.286 - type: map_at_10 value: 30.61 - type: map_at_100 value: 31.619999999999997 - type: map_at_1000 value: 31.724999999999998 - type: map_at_20 value: 31.092 - type: map_at_3 value: 27.962999999999997 - type: map_at_5 value: 29.383 - type: mrr_at_1 value: 24.399260628465804 - type: mrr_at_10 value: 32.696065487193025 - type: mrr_at_100 value: 33.58070054530842 - type: mrr_at_1000 value: 33.654777325405774 - type: mrr_at_20 value: 33.11457792065027 - type: mrr_at_3 value: 30.31423290203327 - type: mrr_at_5 value: 31.552680221811464 - type: nauc_map_at_1000_diff1 value: 39.79013314997585 - type: nauc_map_at_1000_max value: 22.27888746840121 - type: nauc_map_at_1000_std value: 4.539318966935818 - type: nauc_map_at_100_diff1 value: 39.776949099430524 - type: nauc_map_at_100_max value: 22.23994720281935 - type: nauc_map_at_100_std value: 4.532364090321059 - type: nauc_map_at_10_diff1 value: 40.214668109177545 - type: nauc_map_at_10_max value: 22.310555615555764 - type: nauc_map_at_10_std value: 3.7456514343350205 - type: nauc_map_at_1_diff1 value: 46.18393596000586 - type: nauc_map_at_1_max value: 24.547598678024556 - type: nauc_map_at_1_std value: 6.0574958769530465 - type: nauc_map_at_20_diff1 value: 39.95037327455914 - type: nauc_map_at_20_max value: 22.369335761495186 - type: nauc_map_at_20_std value: 4.248871676377463 - type: nauc_map_at_3_diff1 value: 40.39031064698702 - type: nauc_map_at_3_max value: 22.036440129422672 - type: nauc_map_at_3_std value: 2.4849784793648846 - type: nauc_map_at_5_diff1 value: 40.55531780202422 - type: nauc_map_at_5_max value: 22.34099868910038 - type: nauc_map_at_5_std value: 3.989437759311683 - type: nauc_mrr_at_1000_diff1 value: 38.22864890501086 - type: nauc_mrr_at_1000_max value: 22.196145770688915 - type: nauc_mrr_at_1000_std value: 6.366087881052758 - type: nauc_mrr_at_100_diff1 value: 38.19684329027937 - type: nauc_mrr_at_100_max value: 22.17259263583887 - type: nauc_mrr_at_100_std value: 6.3579191826046895 - type: nauc_mrr_at_10_diff1 value: 38.50520505165495 - type: nauc_mrr_at_10_max value: 22.14145550999763 - type: nauc_mrr_at_10_std value: 5.87670477461074 - type: nauc_mrr_at_1_diff1 value: 43.580238226066754 - type: nauc_mrr_at_1_max value: 25.37631028483947 - type: nauc_mrr_at_1_std value: 8.27700367711168 - type: nauc_mrr_at_20_diff1 value: 38.301149084550985 - type: nauc_mrr_at_20_max value: 22.237002751026584 - type: nauc_mrr_at_20_std value: 6.157632931853065 - type: nauc_mrr_at_3_diff1 value: 38.40064989443 - type: nauc_mrr_at_3_max value: 22.300592015957253 - type: nauc_mrr_at_3_std value: 5.111142119521902 - type: nauc_mrr_at_5_diff1 value: 38.74181914377854 - type: nauc_mrr_at_5_max value: 22.25441111952184 - type: nauc_mrr_at_5_std value: 6.22876437673998 - type: nauc_ndcg_at_1000_diff1 value: 36.69736142976795 - type: nauc_ndcg_at_1000_max value: 21.867116284783787 - type: nauc_ndcg_at_1000_std value: 7.265926771096148 - type: nauc_ndcg_at_100_diff1 value: 36.09322471126019 - type: nauc_ndcg_at_100_max value: 21.11550289992875 - type: nauc_ndcg_at_100_std value: 7.040857596769399 - type: nauc_ndcg_at_10_diff1 value: 38.066185877266406 - type: nauc_ndcg_at_10_max value: 21.406313151333396 - type: nauc_ndcg_at_10_std value: 3.714388060329858 - type: nauc_ndcg_at_1_diff1 value: 43.580238226066754 - type: nauc_ndcg_at_1_max value: 25.37631028483947 - type: nauc_ndcg_at_1_std value: 8.27700367711168 - type: nauc_ndcg_at_20_diff1 value: 37.176737325196655 - type: nauc_ndcg_at_20_max value: 21.605872861888944 - type: nauc_ndcg_at_20_std value: 5.139273672061484 - type: nauc_ndcg_at_3_diff1 value: 37.99865829973418 - type: nauc_ndcg_at_3_max value: 21.628352451265933 - type: nauc_ndcg_at_3_std value: 2.5403484884659906 - type: nauc_ndcg_at_5_diff1 value: 38.68827688198417 - type: nauc_ndcg_at_5_max value: 21.766119634697375 - type: nauc_ndcg_at_5_std value: 4.663477639905768 - type: nauc_precision_at_1000_diff1 value: -24.32404164272638 - type: nauc_precision_at_1000_max value: -0.1920006879032294 - type: nauc_precision_at_1000_std value: 3.8459453302163835 - type: nauc_precision_at_100_diff1 value: 0.0961190193116701 - type: nauc_precision_at_100_max value: 10.432470527841613 - type: nauc_precision_at_100_std value: 19.51298317615412 - type: nauc_precision_at_10_diff1 value: 24.865309916077123 - type: nauc_precision_at_10_max value: 19.106193444839885 - type: nauc_precision_at_10_std value: 6.1319125503229985 - type: nauc_precision_at_1_diff1 value: 43.580238226066754 - type: nauc_precision_at_1_max value: 25.37631028483947 - type: nauc_precision_at_1_std value: 8.27700367711168 - type: nauc_precision_at_20_diff1 value: 17.152528821707108 - type: nauc_precision_at_20_max value: 18.550074587326083 - type: nauc_precision_at_20_std value: 12.414087853840773 - type: nauc_precision_at_3_diff1 value: 29.793753328467677 - type: nauc_precision_at_3_max value: 18.856628740486958 - type: nauc_precision_at_3_std value: 1.7490040552720874 - type: nauc_precision_at_5_diff1 value: 27.95189102052665 - type: nauc_precision_at_5_max value: 20.089236844488443 - type: nauc_precision_at_5_std value: 8.272526795799227 - type: nauc_recall_at_1000_diff1 value: 13.869138770344335 - type: nauc_recall_at_1000_max value: 25.76264057259768 - type: nauc_recall_at_1000_std value: 42.620945012763244 - type: nauc_recall_at_100_diff1 value: 18.954723626828734 - type: nauc_recall_at_100_max value: 15.591123917397793 - type: nauc_recall_at_100_std value: 18.872204747720037 - type: nauc_recall_at_10_diff1 value: 32.50173111514971 - type: nauc_recall_at_10_max value: 18.335922588632688 - type: nauc_recall_at_10_std value: 1.6231924423632595 - type: nauc_recall_at_1_diff1 value: 46.18393596000586 - type: nauc_recall_at_1_max value: 24.547598678024556 - type: nauc_recall_at_1_std value: 6.0574958769530465 - type: nauc_recall_at_20_diff1 value: 29.101695015438395 - type: nauc_recall_at_20_max value: 18.63912055487345 - type: nauc_recall_at_20_std value: 6.064314698688468 - type: nauc_recall_at_3_diff1 value: 33.83121888715772 - type: nauc_recall_at_3_max value: 19.258419406401263 - type: nauc_recall_at_3_std value: -0.9541791506796478 - type: nauc_recall_at_5_diff1 value: 34.75197279898117 - type: nauc_recall_at_5_max value: 19.704512261533242 - type: nauc_recall_at_5_std value: 4.482729218598009 - type: ndcg_at_1 value: 24.399 - type: ndcg_at_10 value: 35.555 - type: ndcg_at_100 value: 40.723 - type: ndcg_at_1000 value: 43.155 - type: ndcg_at_20 value: 37.141999999999996 - type: ndcg_at_3 value: 30.45 - type: ndcg_at_5 value: 32.749 - type: precision_at_1 value: 24.399 - type: precision_at_10 value: 5.601 - type: precision_at_100 value: 0.8909999999999999 - type: precision_at_1000 value: 0.121 - type: precision_at_20 value: 3.216 - type: precision_at_3 value: 12.939 - type: precision_at_5 value: 9.168 - type: recall_at_1 value: 22.286 - type: recall_at_10 value: 48.925000000000004 - type: recall_at_100 value: 72.791 - type: recall_at_1000 value: 90.69 - type: recall_at_20 value: 54.649 - type: recall_at_3 value: 35.022 - type: recall_at_5 value: 40.579 - task: type: Retrieval dataset: name: MTEB ClimateFEVER (default) type: mteb/climate-fever config: default split: test revision: 47f2ac6acb640fc46020b02a5b59fdda04d39380 metrics: - type: main_score value: 41.177 - type: map_at_1 value: 17.376 - type: map_at_10 value: 30.705 - type: map_at_100 value: 33.145 - type: map_at_1000 value: 33.324 - type: map_at_20 value: 32.129999999999995 - type: map_at_3 value: 25.352000000000004 - type: map_at_5 value: 28.23 - type: mrr_at_1 value: 41.56351791530945 - type: mrr_at_10 value: 53.54648156765427 - type: mrr_at_100 value: 54.16634559859816 - type: mrr_at_1000 value: 54.18645395732765 - type: mrr_at_20 value: 53.9378827028433 - type: mrr_at_3 value: 50.228013029316024 - type: mrr_at_5 value: 52.172638436482174 - type: nauc_map_at_1000_diff1 value: 6.876082348162679 - type: nauc_map_at_1000_max value: 21.009396900572714 - type: nauc_map_at_1000_std value: 14.066430895753937 - type: nauc_map_at_100_diff1 value: 6.850348439065698 - type: nauc_map_at_100_max value: 21.00364553924676 - type: nauc_map_at_100_std value: 14.059870647686076 - type: nauc_map_at_10_diff1 value: 6.509819157572225 - type: nauc_map_at_10_max value: 20.8065690550504 - type: nauc_map_at_10_std value: 12.562768638969086 - type: nauc_map_at_1_diff1 value: 19.113985692043915 - type: nauc_map_at_1_max value: 27.403489479561337 - type: nauc_map_at_1_std value: 8.997280354530837 - type: nauc_map_at_20_diff1 value: 6.689209935271891 - type: nauc_map_at_20_max value: 20.829284453048967 - type: nauc_map_at_20_std value: 13.537098219731128 - type: nauc_map_at_3_diff1 value: 8.354071849010772 - type: nauc_map_at_3_max value: 21.39794707841315 - type: nauc_map_at_3_std value: 10.16825293444317 - type: nauc_map_at_5_diff1 value: 6.353792564160103 - type: nauc_map_at_5_max value: 20.654610018600735 - type: nauc_map_at_5_std value: 11.51720666348388 - type: nauc_mrr_at_1000_diff1 value: 18.719626503061086 - type: nauc_mrr_at_1000_max value: 25.382297708144915 - type: nauc_mrr_at_1000_std value: 20.795619918235513 - type: nauc_mrr_at_100_diff1 value: 18.707844253612848 - type: nauc_mrr_at_100_max value: 25.37308894691589 - type: nauc_mrr_at_100_std value: 20.792369663110737 - type: nauc_mrr_at_10_diff1 value: 18.599552029091104 - type: nauc_mrr_at_10_max value: 25.27052175696751 - type: nauc_mrr_at_10_std value: 20.780213374556904 - type: nauc_mrr_at_1_diff1 value: 24.986463675582733 - type: nauc_mrr_at_1_max value: 28.633615906622467 - type: nauc_mrr_at_1_std value: 18.935003813583457 - type: nauc_mrr_at_20_diff1 value: 18.58009831602654 - type: nauc_mrr_at_20_max value: 25.309342060502825 - type: nauc_mrr_at_20_std value: 20.811933813239104 - type: nauc_mrr_at_3_diff1 value: 19.03652325617102 - type: nauc_mrr_at_3_max value: 25.590424434633995 - type: nauc_mrr_at_3_std value: 20.672321139371263 - type: nauc_mrr_at_5_diff1 value: 18.62484399593036 - type: nauc_mrr_at_5_max value: 25.69914791020157 - type: nauc_mrr_at_5_std value: 20.85655370414309 - type: nauc_ndcg_at_1000_diff1 value: 8.54695673292356 - type: nauc_ndcg_at_1000_max value: 20.965191922952513 - type: nauc_ndcg_at_1000_std value: 18.638066252011978 - type: nauc_ndcg_at_100_diff1 value: 8.031774449316728 - type: nauc_ndcg_at_100_max value: 21.075278652222494 - type: nauc_ndcg_at_100_std value: 19.202919369605972 - type: nauc_ndcg_at_10_diff1 value: 6.857083069946808 - type: nauc_ndcg_at_10_max value: 20.253829678610604 - type: nauc_ndcg_at_10_std value: 15.456896398668595 - type: nauc_ndcg_at_1_diff1 value: 24.986463675582733 - type: nauc_ndcg_at_1_max value: 28.633615906622467 - type: nauc_ndcg_at_1_std value: 18.935003813583457 - type: nauc_ndcg_at_20_diff1 value: 7.310618350530157 - type: nauc_ndcg_at_20_max value: 20.48058063251671 - type: nauc_ndcg_at_20_std value: 17.35126095861103 - type: nauc_ndcg_at_3_diff1 value: 10.284697710828992 - type: nauc_ndcg_at_3_max value: 21.404564460904535 - type: nauc_ndcg_at_3_std value: 13.811528596529799 - type: nauc_ndcg_at_5_diff1 value: 6.932072809009071 - type: nauc_ndcg_at_5_max value: 20.648949990060657 - type: nauc_ndcg_at_5_std value: 14.368751919376846 - type: nauc_precision_at_1000_diff1 value: -1.4140589422343832 - type: nauc_precision_at_1000_max value: -6.6374826556613264 - type: nauc_precision_at_1000_std value: 11.116149167404775 - type: nauc_precision_at_100_diff1 value: -0.5816105386152639 - type: nauc_precision_at_100_max value: 1.2367532155168361 - type: nauc_precision_at_100_std value: 20.01762008226351 - type: nauc_precision_at_10_diff1 value: -1.8634971794747164 - type: nauc_precision_at_10_max value: 6.8960226644416185 - type: nauc_precision_at_10_std value: 17.20121919885631 - type: nauc_precision_at_1_diff1 value: 24.986463675582733 - type: nauc_precision_at_1_max value: 28.633615906622467 - type: nauc_precision_at_1_std value: 18.935003813583457 - type: nauc_precision_at_20_diff1 value: -1.4459597575880887 - type: nauc_precision_at_20_max value: 5.307806932575533 - type: nauc_precision_at_20_std value: 19.451800377499655 - type: nauc_precision_at_3_diff1 value: 4.236106307523834 - type: nauc_precision_at_3_max value: 14.046883704229765 - type: nauc_precision_at_3_std value: 17.800580068504328 - type: nauc_precision_at_5_diff1 value: -1.3650327582096584 - type: nauc_precision_at_5_max value: 10.207588037756324 - type: nauc_precision_at_5_std value: 17.342725667697678 - type: nauc_recall_at_1000_diff1 value: -0.3456913485138751 - type: nauc_recall_at_1000_max value: 9.035999568091443 - type: nauc_recall_at_1000_std value: 24.89435133186522 - type: nauc_recall_at_100_diff1 value: -0.4515116177152527 - type: nauc_recall_at_100_max value: 13.308695449140274 - type: nauc_recall_at_100_std value: 24.08184104676165 - type: nauc_recall_at_10_diff1 value: -1.7208221232376235 - type: nauc_recall_at_10_max value: 13.184289213175079 - type: nauc_recall_at_10_std value: 12.654581726678604 - type: nauc_recall_at_1_diff1 value: 19.113985692043915 - type: nauc_recall_at_1_max value: 27.403489479561337 - type: nauc_recall_at_1_std value: 8.997280354530837 - type: nauc_recall_at_20_diff1 value: -0.7023429202973307 - type: nauc_recall_at_20_max value: 12.830137977471596 - type: nauc_recall_at_20_std value: 16.37670340447336 - type: nauc_recall_at_3_diff1 value: 2.8972253611264143 - type: nauc_recall_at_3_max value: 16.779165952414292 - type: nauc_recall_at_3_std value: 9.121837904207856 - type: nauc_recall_at_5_diff1 value: -1.2895779049988085 - type: nauc_recall_at_5_max value: 14.974218341119162 - type: nauc_recall_at_5_std value: 11.278321881932376 - type: ndcg_at_1 value: 41.564 - type: ndcg_at_10 value: 41.177 - type: ndcg_at_100 value: 49.036 - type: ndcg_at_1000 value: 51.864 - type: ndcg_at_20 value: 44.535000000000004 - type: ndcg_at_3 value: 34.183 - type: ndcg_at_5 value: 36.636 - type: precision_at_1 value: 41.564 - type: precision_at_10 value: 12.886000000000001 - type: precision_at_100 value: 2.145 - type: precision_at_1000 value: 0.268 - type: precision_at_20 value: 7.922 - type: precision_at_3 value: 25.668000000000003 - type: precision_at_5 value: 19.713 - type: recall_at_1 value: 17.376 - type: recall_at_10 value: 48.116 - type: recall_at_100 value: 73.983 - type: recall_at_1000 value: 89.557 - type: recall_at_20 value: 57.376000000000005 - type: recall_at_3 value: 30.624000000000002 - type: recall_at_5 value: 38.072 - task: type: Retrieval dataset: name: MTEB DBPedia (default) type: mteb/dbpedia config: default split: test revision: c0f706b76e590d620bd6618b3ca8efdd34e2d659 metrics: - type: main_score value: 48.979 - type: map_at_1 value: 9.858 - type: map_at_10 value: 22.772000000000002 - type: map_at_100 value: 32.067 - type: map_at_1000 value: 33.789 - type: map_at_20 value: 26.6 - type: map_at_3 value: 15.817999999999998 - type: map_at_5 value: 18.88 - type: mrr_at_1 value: 75.75 - type: mrr_at_10 value: 81.4422619047619 - type: mrr_at_100 value: 81.66908663880871 - type: mrr_at_1000 value: 81.67603961133557 - type: mrr_at_20 value: 81.58319354256854 - type: mrr_at_3 value: 79.91666666666667 - type: mrr_at_5 value: 81.05416666666666 - type: nauc_map_at_1000_diff1 value: 23.991232076620477 - type: nauc_map_at_1000_max value: 35.287316450717924 - type: nauc_map_at_1000_std value: 19.352326951207623 - type: nauc_map_at_100_diff1 value: 24.340309156988095 - type: nauc_map_at_100_max value: 35.68099330475215 - type: nauc_map_at_100_std value: 17.175838739585252 - type: nauc_map_at_10_diff1 value: 21.561923328790826 - type: nauc_map_at_10_max value: 30.221314679896555 - type: nauc_map_at_10_std value: -5.171508970583829 - type: nauc_map_at_1_diff1 value: 27.8980502688949 - type: nauc_map_at_1_max value: 18.011347877902235 - type: nauc_map_at_1_std value: -21.8828654183319 - type: nauc_map_at_20_diff1 value: 21.97593079179927 - type: nauc_map_at_20_max value: 33.1027436090741 - type: nauc_map_at_20_std value: 3.8376582930602887 - type: nauc_map_at_3_diff1 value: 23.696811666749362 - type: nauc_map_at_3_max value: 25.004984475522406 - type: nauc_map_at_3_std value: -16.036281146384134 - type: nauc_map_at_5_diff1 value: 22.303297302695672 - type: nauc_map_at_5_max value: 25.908488411484537 - type: nauc_map_at_5_std value: -12.727467597748399 - type: nauc_mrr_at_1000_diff1 value: 50.807506669263105 - type: nauc_mrr_at_1000_max value: 59.521138888646895 - type: nauc_mrr_at_1000_std value: 39.72453658171713 - type: nauc_mrr_at_100_diff1 value: 50.809052816882414 - type: nauc_mrr_at_100_max value: 59.52443036190528 - type: nauc_mrr_at_100_std value: 39.71360790190832 - type: nauc_mrr_at_10_diff1 value: 50.71551464513347 - type: nauc_mrr_at_10_max value: 59.46887584854914 - type: nauc_mrr_at_10_std value: 39.720073174909146 - type: nauc_mrr_at_1_diff1 value: 51.23431960913661 - type: nauc_mrr_at_1_max value: 59.02477220193181 - type: nauc_mrr_at_1_std value: 37.613094567706604 - type: nauc_mrr_at_20_diff1 value: 50.68567900468689 - type: nauc_mrr_at_20_max value: 59.398702247575116 - type: nauc_mrr_at_20_std value: 39.84349342123071 - type: nauc_mrr_at_3_diff1 value: 50.84159182980731 - type: nauc_mrr_at_3_max value: 59.586303879639814 - type: nauc_mrr_at_3_std value: 39.115703986532054 - type: nauc_mrr_at_5_diff1 value: 50.9427075304326 - type: nauc_mrr_at_5_max value: 59.9197314639652 - type: nauc_mrr_at_5_std value: 40.03939021575725 - type: nauc_ndcg_at_1000_diff1 value: 35.299374382112134 - type: nauc_ndcg_at_1000_max value: 42.17483524995039 - type: nauc_ndcg_at_1000_std value: 36.65033986688723 - type: nauc_ndcg_at_100_diff1 value: 34.44823939199226 - type: nauc_ndcg_at_100_max value: 41.7528959441004 - type: nauc_ndcg_at_100_std value: 28.72365119802961 - type: nauc_ndcg_at_10_diff1 value: 29.32293547048091 - type: nauc_ndcg_at_10_max value: 40.101679400646006 - type: nauc_ndcg_at_10_std value: 26.5721071370353 - type: nauc_ndcg_at_1_diff1 value: 48.319456575299284 - type: nauc_ndcg_at_1_max value: 48.27377641677222 - type: nauc_ndcg_at_1_std value: 29.76971701564757 - type: nauc_ndcg_at_20_diff1 value: 30.927032015266835 - type: nauc_ndcg_at_20_max value: 40.52043580178855 - type: nauc_ndcg_at_20_std value: 25.197926348678955 - type: nauc_ndcg_at_3_diff1 value: 33.082418428993115 - type: nauc_ndcg_at_3_max value: 40.62252050374572 - type: nauc_ndcg_at_3_std value: 28.113380979394726 - type: nauc_ndcg_at_5_diff1 value: 29.635117682340617 - type: nauc_ndcg_at_5_max value: 38.11353464984394 - type: nauc_ndcg_at_5_std value: 28.33324261545152 - type: nauc_precision_at_1000_diff1 value: -18.548687962963978 - type: nauc_precision_at_1000_max value: -14.491062706051878 - type: nauc_precision_at_1000_std value: 10.681709294585238 - type: nauc_precision_at_100_diff1 value: -1.688856131371092 - type: nauc_precision_at_100_max value: 5.481319501702683 - type: nauc_precision_at_100_std value: 39.979879645237446 - type: nauc_precision_at_10_diff1 value: 0.576840213887176 - type: nauc_precision_at_10_max value: 18.614962845466955 - type: nauc_precision_at_10_std value: 42.024684223351464 - type: nauc_precision_at_1_diff1 value: 51.23431960913661 - type: nauc_precision_at_1_max value: 59.02477220193181 - type: nauc_precision_at_1_std value: 37.613094567706604 - type: nauc_precision_at_20_diff1 value: 1.3707715784045262 - type: nauc_precision_at_20_max value: 14.922028634512083 - type: nauc_precision_at_20_std value: 44.76530134675204 - type: nauc_precision_at_3_diff1 value: 13.094243395849992 - type: nauc_precision_at_3_max value: 29.850584449565037 - type: nauc_precision_at_3_std value: 35.77371986318991 - type: nauc_precision_at_5_diff1 value: 6.798339179999441 - type: nauc_precision_at_5_max value: 23.08541604839939 - type: nauc_precision_at_5_std value: 40.28922731098164 - type: nauc_recall_at_1000_diff1 value: 27.24738341174725 - type: nauc_recall_at_1000_max value: 31.09981332493123 - type: nauc_recall_at_1000_std value: 41.96422474881881 - type: nauc_recall_at_100_diff1 value: 24.922315595458294 - type: nauc_recall_at_100_max value: 32.53690673184911 - type: nauc_recall_at_100_std value: 23.02548177144121 - type: nauc_recall_at_10_diff1 value: 15.873395525740868 - type: nauc_recall_at_10_max value: 23.963191643746132 - type: nauc_recall_at_10_std value: -7.2368622521479296 - type: nauc_recall_at_1_diff1 value: 27.8980502688949 - type: nauc_recall_at_1_max value: 18.011347877902235 - type: nauc_recall_at_1_std value: -21.8828654183319 - type: nauc_recall_at_20_diff1 value: 17.63321564134115 - type: nauc_recall_at_20_max value: 27.284001947728797 - type: nauc_recall_at_20_std value: 2.3851101283717666 - type: nauc_recall_at_3_diff1 value: 21.72291192189032 - type: nauc_recall_at_3_max value: 23.109590882141113 - type: nauc_recall_at_3_std value: -16.34348495895044 - type: nauc_recall_at_5_diff1 value: 17.596468253564954 - type: nauc_recall_at_5_max value: 20.664891173216603 - type: nauc_recall_at_5_std value: -14.565623699193717 - type: ndcg_at_1 value: 65.125 - type: ndcg_at_10 value: 48.979 - type: ndcg_at_100 value: 52.317 - type: ndcg_at_1000 value: 59.424 - type: ndcg_at_20 value: 47.806 - type: ndcg_at_3 value: 54.032000000000004 - type: ndcg_at_5 value: 51.520999999999994 - type: precision_at_1 value: 75.75 - type: precision_at_10 value: 38.975 - type: precision_at_100 value: 11.848 - type: precision_at_1000 value: 2.199 - type: precision_at_20 value: 29.387 - type: precision_at_3 value: 57.333 - type: precision_at_5 value: 50.0 - type: recall_at_1 value: 9.858 - type: recall_at_10 value: 28.061999999999998 - type: recall_at_100 value: 56.413000000000004 - type: recall_at_1000 value: 79.963 - type: recall_at_20 value: 36.161 - type: recall_at_3 value: 16.631 - type: recall_at_5 value: 21.363 - task: type: Classification dataset: name: MTEB EmotionClassification (default) type: mteb/emotion config: default split: test revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37 metrics: - type: accuracy value: 79.985 - type: f1 value: 74.58481640194556 - type: f1_weighted value: 80.85307620086522 - type: main_score value: 79.985 - task: type: Retrieval dataset: name: MTEB FEVER (default) type: mteb/fever config: default split: test revision: bea83ef9e8fb933d90a2f1d5515737465d613e12 metrics: - type: main_score value: 88.228 - type: map_at_1 value: 75.053 - type: map_at_10 value: 84.631 - type: map_at_100 value: 84.832 - type: map_at_1000 value: 84.844 - type: map_at_20 value: 84.756 - type: map_at_3 value: 83.553 - type: map_at_5 value: 84.273 - type: mrr_at_1 value: 80.993099309931 - type: mrr_at_10 value: 88.46116754532582 - type: mrr_at_100 value: 88.51308773245357 - type: mrr_at_1000 value: 88.51396221984112 - type: mrr_at_20 value: 88.49642168590128 - type: mrr_at_3 value: 87.86128612861275 - type: mrr_at_5 value: 88.30083008300812 - type: nauc_map_at_1000_diff1 value: 53.97823649031981 - type: nauc_map_at_1000_max value: 26.11243091917843 - type: nauc_map_at_1000_std value: -10.057644234421986 - type: nauc_map_at_100_diff1 value: 53.925080266687445 - type: nauc_map_at_100_max value: 26.074046044483406 - type: nauc_map_at_100_std value: -10.057139918091936 - type: nauc_map_at_10_diff1 value: 53.378232678228535 - type: nauc_map_at_10_max value: 25.583629956942904 - type: nauc_map_at_10_std value: -10.296034633396092 - type: nauc_map_at_1_diff1 value: 60.507796141511605 - type: nauc_map_at_1_max value: 24.81979211893891 - type: nauc_map_at_1_std value: -15.864717081534302 - type: nauc_map_at_20_diff1 value: 53.712573269726484 - type: nauc_map_at_20_max value: 25.870196380003335 - type: nauc_map_at_20_std value: -10.139248046597455 - type: nauc_map_at_3_diff1 value: 53.261264809399556 - type: nauc_map_at_3_max value: 25.65803011606916 - type: nauc_map_at_3_std value: -10.953616682218243 - type: nauc_map_at_5_diff1 value: 53.17212766431546 - type: nauc_map_at_5_max value: 25.60582034909538 - type: nauc_map_at_5_std value: -10.32613724902313 - type: nauc_mrr_at_1000_diff1 value: 70.38955167949939 - type: nauc_mrr_at_1000_max value: 39.821515037282204 - type: nauc_mrr_at_1000_std value: -9.98013185324074 - type: nauc_mrr_at_100_diff1 value: 70.38352452325266 - type: nauc_mrr_at_100_max value: 39.82466363867733 - type: nauc_mrr_at_100_std value: -9.976145831114493 - type: nauc_mrr_at_10_diff1 value: 70.26683508867457 - type: nauc_mrr_at_10_max value: 39.80122496712571 - type: nauc_mrr_at_10_std value: -9.909384325865775 - type: nauc_mrr_at_1_diff1 value: 73.24890171347613 - type: nauc_mrr_at_1_max value: 37.367459553642426 - type: nauc_mrr_at_1_std value: -13.316391532791135 - type: nauc_mrr_at_20_diff1 value: 70.34500637714407 - type: nauc_mrr_at_20_max value: 39.84118580511733 - type: nauc_mrr_at_20_std value: -9.920771311393942 - type: nauc_mrr_at_3_diff1 value: 70.04420618345499 - type: nauc_mrr_at_3_max value: 40.33885175872482 - type: nauc_mrr_at_3_std value: -9.2308606747524 - type: nauc_mrr_at_5_diff1 value: 70.23298852823912 - type: nauc_mrr_at_5_max value: 40.28613289657475 - type: nauc_mrr_at_5_std value: -9.408644815171415 - type: nauc_ndcg_at_1000_diff1 value: 56.14884407654613 - type: nauc_ndcg_at_1000_max value: 29.027269391217793 - type: nauc_ndcg_at_1000_std value: -8.185655036370417 - type: nauc_ndcg_at_100_diff1 value: 54.898228209830854 - type: nauc_ndcg_at_100_max value: 28.23127072967732 - type: nauc_ndcg_at_100_std value: -7.937951960666996 - type: nauc_ndcg_at_10_diff1 value: 52.76884326536276 - type: nauc_ndcg_at_10_max value: 26.501133559532004 - type: nauc_ndcg_at_10_std value: -8.561291306720568 - type: nauc_ndcg_at_1_diff1 value: 73.24890171347613 - type: nauc_ndcg_at_1_max value: 37.367459553642426 - type: nauc_ndcg_at_1_std value: -13.316391532791135 - type: nauc_ndcg_at_20_diff1 value: 53.782879241534154 - type: nauc_ndcg_at_20_max value: 27.344714620733146 - type: nauc_ndcg_at_20_std value: -8.174365511016143 - type: nauc_ndcg_at_3_diff1 value: 54.07748391367295 - type: nauc_ndcg_at_3_max value: 28.740769448822867 - type: nauc_ndcg_at_3_std value: -8.800638719106981 - type: nauc_ndcg_at_5_diff1 value: 52.91102194973326 - type: nauc_ndcg_at_5_max value: 27.35297204098582 - type: nauc_ndcg_at_5_std value: -8.202780538104845 - type: nauc_precision_at_1000_diff1 value: -6.462960135986346 - type: nauc_precision_at_1000_max value: 12.759892798322381 - type: nauc_precision_at_1000_std value: 17.830413795603956 - type: nauc_precision_at_100_diff1 value: -10.714161244793623 - type: nauc_precision_at_100_max value: 10.80916133379338 - type: nauc_precision_at_100_std value: 21.01280694690889 - type: nauc_precision_at_10_diff1 value: -12.867253218059915 - type: nauc_precision_at_10_max value: 9.575643543429718 - type: nauc_precision_at_10_std value: 21.405171955259224 - type: nauc_precision_at_1_diff1 value: 73.24890171347613 - type: nauc_precision_at_1_max value: 37.367459553642426 - type: nauc_precision_at_1_std value: -13.316391532791135 - type: nauc_precision_at_20_diff1 value: -11.766460335141424 - type: nauc_precision_at_20_max value: 10.17190973145006 - type: nauc_precision_at_20_std value: 21.752924700590835 - type: nauc_precision_at_3_diff1 value: 5.241669513189873 - type: nauc_precision_at_3_max value: 21.722890037760354 - type: nauc_precision_at_3_std value: 16.83232605784222 - type: nauc_precision_at_5_diff1 value: -6.750151592516413 - type: nauc_precision_at_5_max value: 15.059744329415048 - type: nauc_precision_at_5_std value: 21.831836531443653 - type: nauc_recall_at_1000_diff1 value: 8.852828649246417 - type: nauc_recall_at_1000_max value: 6.683830914994345 - type: nauc_recall_at_1000_std value: 37.66593889403836 - type: nauc_recall_at_100_diff1 value: -2.4986179820673344 - type: nauc_recall_at_100_max value: -1.230471742842536 - type: nauc_recall_at_100_std value: 22.724612835383482 - type: nauc_recall_at_10_diff1 value: 8.921193487520886 - type: nauc_recall_at_10_max value: 1.4012350766088484 - type: nauc_recall_at_10_std value: 2.9284367419689037 - type: nauc_recall_at_1_diff1 value: 60.507796141511605 - type: nauc_recall_at_1_max value: 24.81979211893891 - type: nauc_recall_at_1_std value: -15.864717081534302 - type: nauc_recall_at_20_diff1 value: 6.778598529994739 - type: nauc_recall_at_20_max value: 1.9108915219621572 - type: nauc_recall_at_20_std value: 9.15003581851989 - type: nauc_recall_at_3_diff1 value: 30.17670764440773 - type: nauc_recall_at_3_max value: 17.769313053478434 - type: nauc_recall_at_3_std value: -2.7848998516990386 - type: nauc_recall_at_5_diff1 value: 19.986644381812553 - type: nauc_recall_at_5_max value: 11.751813635626322 - type: nauc_recall_at_5_std value: 1.6890369172263033 - type: ndcg_at_1 value: 80.99300000000001 - type: ndcg_at_10 value: 88.228 - type: ndcg_at_100 value: 88.897 - type: ndcg_at_1000 value: 89.093 - type: ndcg_at_20 value: 88.542 - type: ndcg_at_3 value: 86.687 - type: ndcg_at_5 value: 87.607 - type: precision_at_1 value: 80.99300000000001 - type: precision_at_10 value: 10.707 - type: precision_at_100 value: 1.127 - type: precision_at_1000 value: 0.116 - type: precision_at_20 value: 5.457999999999999 - type: precision_at_3 value: 33.538000000000004 - type: precision_at_5 value: 20.801 - type: recall_at_1 value: 75.053 - type: recall_at_10 value: 95.27799999999999 - type: recall_at_100 value: 97.853 - type: recall_at_1000 value: 99.03800000000001 - type: recall_at_20 value: 96.318 - type: recall_at_3 value: 91.08000000000001 - type: recall_at_5 value: 93.45400000000001 - task: type: Retrieval dataset: name: MTEB FiQA2018 (default) type: mteb/fiqa config: default split: test revision: 27a168819829fe9bcd655c2df245fb19452e8e06 metrics: - type: main_score value: 54.071999999999996 - type: map_at_1 value: 27.345000000000002 - type: map_at_10 value: 45.694 - type: map_at_100 value: 47.949999999999996 - type: map_at_1000 value: 48.093 - type: map_at_20 value: 47.035 - type: map_at_3 value: 40.049 - type: map_at_5 value: 42.92 - type: mrr_at_1 value: 52.160493827160494 - type: mrr_at_10 value: 61.28527336860672 - type: mrr_at_100 value: 61.884625221750596 - type: mrr_at_1000 value: 61.915904963540726 - type: mrr_at_20 value: 61.667012493286734 - type: mrr_at_3 value: 59.10493827160492 - type: mrr_at_5 value: 60.362654320987616 - type: nauc_map_at_1000_diff1 value: 45.38067605515959 - type: nauc_map_at_1000_max value: 37.86197840734717 - type: nauc_map_at_1000_std value: -10.88609599497855 - type: nauc_map_at_100_diff1 value: 45.34384139935809 - type: nauc_map_at_100_max value: 37.79374163212799 - type: nauc_map_at_100_std value: -10.800059266281165 - type: nauc_map_at_10_diff1 value: 45.3913490132632 - type: nauc_map_at_10_max value: 36.79840356578914 - type: nauc_map_at_10_std value: -12.036133364054884 - type: nauc_map_at_1_diff1 value: 51.66552774401746 - type: nauc_map_at_1_max value: 25.324194752193236 - type: nauc_map_at_1_std value: -14.34697090462958 - type: nauc_map_at_20_diff1 value: 45.27320308873338 - type: nauc_map_at_20_max value: 37.29442746411085 - type: nauc_map_at_20_std value: -11.635204276133472 - type: nauc_map_at_3_diff1 value: 46.88138818586725 - type: nauc_map_at_3_max value: 32.99288436262902 - type: nauc_map_at_3_std value: -13.639274978165444 - type: nauc_map_at_5_diff1 value: 45.76135530895121 - type: nauc_map_at_5_max value: 34.97804527762444 - type: nauc_map_at_5_std value: -12.678346477642899 - type: nauc_mrr_at_1000_diff1 value: 53.864293429447955 - type: nauc_mrr_at_1000_max value: 45.79808916389802 - type: nauc_mrr_at_1000_std value: -9.713381409523494 - type: nauc_mrr_at_100_diff1 value: 53.85134409074757 - type: nauc_mrr_at_100_max value: 45.80389587114905 - type: nauc_mrr_at_100_std value: -9.683169165384212 - type: nauc_mrr_at_10_diff1 value: 53.805490205878 - type: nauc_mrr_at_10_max value: 45.806760270208564 - type: nauc_mrr_at_10_std value: -9.76722195012393 - type: nauc_mrr_at_1_diff1 value: 56.27330361790344 - type: nauc_mrr_at_1_max value: 47.01503122847836 - type: nauc_mrr_at_1_std value: -10.774154484447495 - type: nauc_mrr_at_20_diff1 value: 53.83482468037953 - type: nauc_mrr_at_20_max value: 45.719679695052974 - type: nauc_mrr_at_20_std value: -9.77923533594551 - type: nauc_mrr_at_3_diff1 value: 54.44641861789147 - type: nauc_mrr_at_3_max value: 45.94905694818705 - type: nauc_mrr_at_3_std value: -11.177467065728768 - type: nauc_mrr_at_5_diff1 value: 54.09429588760707 - type: nauc_mrr_at_5_max value: 46.004166041517216 - type: nauc_mrr_at_5_std value: -9.769538819499722 - type: nauc_ndcg_at_1000_diff1 value: 46.80179242198247 - type: nauc_ndcg_at_1000_max value: 40.806989668058186 - type: nauc_ndcg_at_1000_std value: -8.015013067414483 - type: nauc_ndcg_at_100_diff1 value: 46.26031710590574 - type: nauc_ndcg_at_100_max value: 40.2874844490879 - type: nauc_ndcg_at_100_std value: -6.325738537481981 - type: nauc_ndcg_at_10_diff1 value: 46.0597385861321 - type: nauc_ndcg_at_10_max value: 38.12369512757341 - type: nauc_ndcg_at_10_std value: -9.95387894167638 - type: nauc_ndcg_at_1_diff1 value: 56.27330361790344 - type: nauc_ndcg_at_1_max value: 47.01503122847836 - type: nauc_ndcg_at_1_std value: -10.774154484447495 - type: nauc_ndcg_at_20_diff1 value: 46.112983276165046 - type: nauc_ndcg_at_20_max value: 38.60654549021085 - type: nauc_ndcg_at_20_std value: -9.66055049547148 - type: nauc_ndcg_at_3_diff1 value: 46.07426386701122 - type: nauc_ndcg_at_3_max value: 39.30739016101109 - type: nauc_ndcg_at_3_std value: -12.50493736255984 - type: nauc_ndcg_at_5_diff1 value: 45.71298951268576 - type: nauc_ndcg_at_5_max value: 37.27961846995706 - type: nauc_ndcg_at_5_std value: -11.154006989020496 - type: nauc_precision_at_1000_diff1 value: -11.592438042119445 - type: nauc_precision_at_1000_max value: 17.294449668418288 - type: nauc_precision_at_1000_std value: 9.709962161201647 - type: nauc_precision_at_100_diff1 value: -6.0095430176874345 - type: nauc_precision_at_100_max value: 22.901828845166698 - type: nauc_precision_at_100_std value: 14.993379617197682 - type: nauc_precision_at_10_diff1 value: 6.719203274493172 - type: nauc_precision_at_10_max value: 32.512145720381795 - type: nauc_precision_at_10_std value: 5.244187871424349 - type: nauc_precision_at_1_diff1 value: 56.27330361790344 - type: nauc_precision_at_1_max value: 47.01503122847836 - type: nauc_precision_at_1_std value: -10.774154484447495 - type: nauc_precision_at_20_diff1 value: 1.754389508301811 - type: nauc_precision_at_20_max value: 29.02035054956672 - type: nauc_precision_at_20_std value: 8.161759871402037 - type: nauc_precision_at_3_diff1 value: 24.040968725090252 - type: nauc_precision_at_3_max value: 40.10318275587437 - type: nauc_precision_at_3_std value: -3.878413890678057 - type: nauc_precision_at_5_diff1 value: 15.218812798552142 - type: nauc_precision_at_5_max value: 37.25953351705925 - type: nauc_precision_at_5_std value: 0.7155796998283327 - type: nauc_recall_at_1000_diff1 value: 10.583253250637997 - type: nauc_recall_at_1000_max value: -3.5637377831543846 - type: nauc_recall_at_1000_std value: 34.74872993454209 - type: nauc_recall_at_100_diff1 value: 26.680647396718747 - type: nauc_recall_at_100_max value: 25.289227360067045 - type: nauc_recall_at_100_std value: 19.215575737374877 - type: nauc_recall_at_10_diff1 value: 35.49850774071538 - type: nauc_recall_at_10_max value: 27.12975488283297 - type: nauc_recall_at_10_std value: -6.7757139852899995 - type: nauc_recall_at_1_diff1 value: 51.66552774401746 - type: nauc_recall_at_1_max value: 25.324194752193236 - type: nauc_recall_at_1_std value: -14.34697090462958 - type: nauc_recall_at_20_diff1 value: 33.87213110916921 - type: nauc_recall_at_20_max value: 25.15617289177912 - type: nauc_recall_at_20_std value: -6.44141075455468 - type: nauc_recall_at_3_diff1 value: 42.167552979112784 - type: nauc_recall_at_3_max value: 26.47073745859859 - type: nauc_recall_at_3_std value: -13.151450499133057 - type: nauc_recall_at_5_diff1 value: 38.5058386963604 - type: nauc_recall_at_5_max value: 26.128698034399218 - type: nauc_recall_at_5_std value: -8.92423552488776 - type: ndcg_at_1 value: 52.16 - type: ndcg_at_10 value: 54.071999999999996 - type: ndcg_at_100 value: 60.851 - type: ndcg_at_1000 value: 62.907999999999994 - type: ndcg_at_20 value: 57.001000000000005 - type: ndcg_at_3 value: 49.712 - type: ndcg_at_5 value: 50.791 - type: precision_at_1 value: 52.16 - type: precision_at_10 value: 15.062000000000001 - type: precision_at_100 value: 2.218 - type: precision_at_1000 value: 0.258 - type: precision_at_20 value: 8.827 - type: precision_at_3 value: 33.282000000000004 - type: precision_at_5 value: 24.012 - type: recall_at_1 value: 27.345000000000002 - type: recall_at_10 value: 61.846999999999994 - type: recall_at_100 value: 86.125 - type: recall_at_1000 value: 98.13199999999999 - type: recall_at_20 value: 70.545 - type: recall_at_3 value: 45.446 - type: recall_at_5 value: 52.031000000000006 - task: type: Retrieval dataset: name: MTEB HotpotQA (default) type: mteb/hotpotqa config: default split: test revision: ab518f4d6fcca38d87c25209f94beba119d02014 metrics: - type: main_score value: 74.959 - type: map_at_1 value: 39.507 - type: map_at_10 value: 67.368 - type: map_at_100 value: 68.208 - type: map_at_1000 value: 68.258 - type: map_at_20 value: 67.9 - type: map_at_3 value: 63.695 - type: map_at_5 value: 66.069 - type: mrr_at_1 value: 79.01417960837273 - type: mrr_at_10 value: 85.29256294009818 - type: mrr_at_100 value: 85.43598687762321 - type: mrr_at_1000 value: 85.44005185885888 - type: mrr_at_20 value: 85.385908910377 - type: mrr_at_3 value: 84.41368444744523 - type: mrr_at_5 value: 84.9990997074046 - type: nauc_map_at_1000_diff1 value: 21.12170489619495 - type: nauc_map_at_1000_max value: 26.275894183746722 - type: nauc_map_at_1000_std value: -6.270422764724773 - type: nauc_map_at_100_diff1 value: 21.100748412891427 - type: nauc_map_at_100_max value: 26.267357900952376 - type: nauc_map_at_100_std value: -6.244347667315573 - type: nauc_map_at_10_diff1 value: 20.674777133569105 - type: nauc_map_at_10_max value: 26.0464950302885 - type: nauc_map_at_10_std value: -6.879486555235194 - type: nauc_map_at_1_diff1 value: 61.578918111691614 - type: nauc_map_at_1_max value: 42.809228851971554 - type: nauc_map_at_1_std value: -18.693501607160478 - type: nauc_map_at_20_diff1 value: 21.016679127441627 - type: nauc_map_at_20_max value: 26.26493055547197 - type: nauc_map_at_20_std value: -6.348265956664924 - type: nauc_map_at_3_diff1 value: 19.211524514269673 - type: nauc_map_at_3_max value: 25.179630796295072 - type: nauc_map_at_3_std value: -9.469682815051597 - type: nauc_map_at_5_diff1 value: 19.802257269903983 - type: nauc_map_at_5_max value: 25.843065189828675 - type: nauc_map_at_5_std value: -7.6911117288836275 - type: nauc_mrr_at_1000_diff1 value: 60.90611255392621 - type: nauc_mrr_at_1000_max value: 45.28902337460921 - type: nauc_mrr_at_1000_std value: -15.081836800607629 - type: nauc_mrr_at_100_diff1 value: 60.906319613903634 - type: nauc_mrr_at_100_max value: 45.294454122569135 - type: nauc_mrr_at_100_std value: -15.070354934845525 - type: nauc_mrr_at_10_diff1 value: 60.89081258769886 - type: nauc_mrr_at_10_max value: 45.340063090713706 - type: nauc_mrr_at_10_std value: -15.019436328769977 - type: nauc_mrr_at_1_diff1 value: 61.578918111691614 - type: nauc_mrr_at_1_max value: 42.809228851971554 - type: nauc_mrr_at_1_std value: -18.693501607160478 - type: nauc_mrr_at_20_diff1 value: 60.91444288979141 - type: nauc_mrr_at_20_max value: 45.31431373445948 - type: nauc_mrr_at_20_std value: -14.97309014683095 - type: nauc_mrr_at_3_diff1 value: 60.772894031312696 - type: nauc_mrr_at_3_max value: 45.605293386022225 - type: nauc_mrr_at_3_std value: -15.391241831624658 - type: nauc_mrr_at_5_diff1 value: 60.71990183490615 - type: nauc_mrr_at_5_max value: 45.478031078283045 - type: nauc_mrr_at_5_std value: -15.099732959629012 - type: nauc_ndcg_at_1000_diff1 value: 27.86370916809549 - type: nauc_ndcg_at_1000_max value: 29.961195201820917 - type: nauc_ndcg_at_1000_std value: -3.669547648606182 - type: nauc_ndcg_at_100_diff1 value: 27.222363197903203 - type: nauc_ndcg_at_100_max value: 29.83590955603319 - type: nauc_ndcg_at_100_std value: -2.706914023646432 - type: nauc_ndcg_at_10_diff1 value: 25.720275283710905 - type: nauc_ndcg_at_10_max value: 29.099451842124513 - type: nauc_ndcg_at_10_std value: -4.974149196543083 - type: nauc_ndcg_at_1_diff1 value: 61.578918111691614 - type: nauc_ndcg_at_1_max value: 42.809228851971554 - type: nauc_ndcg_at_1_std value: -18.693501607160478 - type: nauc_ndcg_at_20_diff1 value: 26.6414778719889 - type: nauc_ndcg_at_20_max value: 29.7153470420483 - type: nauc_ndcg_at_20_std value: -3.323674164247545 - type: nauc_ndcg_at_3_diff1 value: 23.854705547556676 - type: nauc_ndcg_at_3_max value: 28.16819582399863 - type: nauc_ndcg_at_3_std value: -9.175742937548364 - type: nauc_ndcg_at_5_diff1 value: 24.235289969946336 - type: nauc_ndcg_at_5_max value: 28.837159697000736 - type: nauc_ndcg_at_5_std value: -6.6312641809059825 - type: nauc_precision_at_1000_diff1 value: 15.588021360728687 - type: nauc_precision_at_1000_max value: 22.39953961246837 - type: nauc_precision_at_1000_std value: 47.68406787651948 - type: nauc_precision_at_100_diff1 value: 14.082191847912181 - type: nauc_precision_at_100_max value: 24.398280717374227 - type: nauc_precision_at_100_std value: 29.845964300686106 - type: nauc_precision_at_10_diff1 value: 14.078430107561424 - type: nauc_precision_at_10_max value: 24.03621964514711 - type: nauc_precision_at_10_std value: 6.216273371941104 - type: nauc_precision_at_1_diff1 value: 61.578918111691614 - type: nauc_precision_at_1_max value: 42.809228851971554 - type: nauc_precision_at_1_std value: -18.693501607160478 - type: nauc_precision_at_20_diff1 value: 15.305783955465262 - type: nauc_precision_at_20_max value: 25.331504698917186 - type: nauc_precision_at_20_std value: 14.995465986068544 - type: nauc_precision_at_3_diff1 value: 13.428353704090718 - type: nauc_precision_at_3_max value: 24.235140590205866 - type: nauc_precision_at_3_std value: -5.482186394535428 - type: nauc_precision_at_5_diff1 value: 12.446233438129173 - type: nauc_precision_at_5_max value: 24.572973116392642 - type: nauc_precision_at_5_std value: 0.24277662503234543 - type: nauc_recall_at_1000_diff1 value: 15.588021360729346 - type: nauc_recall_at_1000_max value: 22.399539612468512 - type: nauc_recall_at_1000_std value: 47.684067876519485 - type: nauc_recall_at_100_diff1 value: 14.082191847912085 - type: nauc_recall_at_100_max value: 24.398280717374345 - type: nauc_recall_at_100_std value: 29.845964300686095 - type: nauc_recall_at_10_diff1 value: 14.078430107561557 - type: nauc_recall_at_10_max value: 24.03621964514713 - type: nauc_recall_at_10_std value: 6.216273371941132 - type: nauc_recall_at_1_diff1 value: 61.578918111691614 - type: nauc_recall_at_1_max value: 42.809228851971554 - type: nauc_recall_at_1_std value: -18.693501607160478 - type: nauc_recall_at_20_diff1 value: 15.30578395546536 - type: nauc_recall_at_20_max value: 25.33150469891727 - type: nauc_recall_at_20_std value: 14.995465986068684 - type: nauc_recall_at_3_diff1 value: 13.428353704090698 - type: nauc_recall_at_3_max value: 24.235140590205813 - type: nauc_recall_at_3_std value: -5.482186394535521 - type: nauc_recall_at_5_diff1 value: 12.446233438129164 - type: nauc_recall_at_5_max value: 24.572973116392614 - type: nauc_recall_at_5_std value: 0.242776625032411 - type: ndcg_at_1 value: 79.014 - type: ndcg_at_10 value: 74.959 - type: ndcg_at_100 value: 77.70700000000001 - type: ndcg_at_1000 value: 78.628 - type: ndcg_at_20 value: 76.23400000000001 - type: ndcg_at_3 value: 69.891 - type: ndcg_at_5 value: 72.82600000000001 - type: precision_at_1 value: 79.014 - type: precision_at_10 value: 15.946 - type: precision_at_100 value: 1.806 - type: precision_at_1000 value: 0.193 - type: precision_at_20 value: 8.381 - type: precision_at_3 value: 45.726 - type: precision_at_5 value: 29.75 - type: recall_at_1 value: 39.507 - type: recall_at_10 value: 79.73 - type: recall_at_100 value: 90.28399999999999 - type: recall_at_1000 value: 96.327 - type: recall_at_20 value: 83.815 - type: recall_at_3 value: 68.589 - type: recall_at_5 value: 74.375 - task: type: Classification dataset: name: MTEB ImdbClassification (default) type: mteb/imdb config: default split: test revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7 metrics: - type: accuracy value: 96.6444 - type: ap value: 94.71353946976426 - type: ap_weighted value: 94.71353946976426 - type: f1 value: 96.64368622221421 - type: f1_weighted value: 96.64368622221419 - type: main_score value: 96.6444 - task: type: Retrieval dataset: name: MTEB MSMARCO (default) type: mteb/msmarco config: default split: dev revision: c5a29a104738b98a9e76336939199e264163d4a0 metrics: - type: main_score value: 42.545 - type: map_at_1 value: 22.655 - type: map_at_10 value: 35.467999999999996 - type: map_at_100 value: 36.652 - type: map_at_1000 value: 36.692 - type: map_at_20 value: 36.236000000000004 - type: map_at_3 value: 31.485000000000003 - type: map_at_5 value: 33.908 - type: mrr_at_1 value: 23.280802292263612 - type: mrr_at_10 value: 36.02329217264742 - type: mrr_at_100 value: 37.148118537446415 - type: mrr_at_1000 value: 37.183801105059956 - type: mrr_at_20 value: 36.76525675340281 - type: mrr_at_3 value: 32.096466093600654 - type: mrr_at_5 value: 34.50334288443163 - type: nauc_map_at_1000_diff1 value: 34.520324529857184 - type: nauc_map_at_1000_max value: 35.326534835022514 - type: nauc_map_at_1000_std value: -21.366160566488187 - type: nauc_map_at_100_diff1 value: 34.51815749165448 - type: nauc_map_at_100_max value: 35.36490672794807 - type: nauc_map_at_100_std value: -21.34319223709314 - type: nauc_map_at_10_diff1 value: 34.36733390350321 - type: nauc_map_at_10_max value: 35.47907368100861 - type: nauc_map_at_10_std value: -21.932334599571735 - type: nauc_map_at_1_diff1 value: 37.89554066876773 - type: nauc_map_at_1_max value: 28.579792597905413 - type: nauc_map_at_1_std value: -20.51606339206856 - type: nauc_map_at_20_diff1 value: 34.51926497566516 - type: nauc_map_at_20_max value: 35.497148638709895 - type: nauc_map_at_20_std value: -21.595925239714767 - type: nauc_map_at_3_diff1 value: 34.64924634604746 - type: nauc_map_at_3_max value: 33.298757220805754 - type: nauc_map_at_3_std value: -22.44092979514115 - type: nauc_map_at_5_diff1 value: 34.52262452267762 - type: nauc_map_at_5_max value: 34.993794904126 - type: nauc_map_at_5_std value: -22.19799323514771 - type: nauc_mrr_at_1000_diff1 value: 34.30028152962552 - type: nauc_mrr_at_1000_max value: 34.84294030005338 - type: nauc_mrr_at_1000_std value: -21.3040159303398 - type: nauc_mrr_at_100_diff1 value: 34.29714922716057 - type: nauc_mrr_at_100_max value: 34.8773691257525 - type: nauc_mrr_at_100_std value: -21.280800887086606 - type: nauc_mrr_at_10_diff1 value: 34.141133687651255 - type: nauc_mrr_at_10_max value: 34.97057209823848 - type: nauc_mrr_at_10_std value: -21.82443447975521 - type: nauc_mrr_at_1_diff1 value: 37.68273289251851 - type: nauc_mrr_at_1_max value: 28.375793374298752 - type: nauc_mrr_at_1_std value: -20.548630760150132 - type: nauc_mrr_at_20_diff1 value: 34.29297087665669 - type: nauc_mrr_at_20_max value: 34.99361503254817 - type: nauc_mrr_at_20_std value: -21.492481020546688 - type: nauc_mrr_at_3_diff1 value: 34.46337545862703 - type: nauc_mrr_at_3_max value: 32.91269289394109 - type: nauc_mrr_at_3_std value: -22.400479840328636 - type: nauc_mrr_at_5_diff1 value: 34.28655737221164 - type: nauc_mrr_at_5_max value: 34.57504983292885 - type: nauc_mrr_at_5_std value: -22.11521048383527 - type: nauc_ndcg_at_1000_diff1 value: 33.62874580335025 - type: nauc_ndcg_at_1000_max value: 37.1872988379906 - type: nauc_ndcg_at_1000_std value: -19.60332451143694 - type: nauc_ndcg_at_100_diff1 value: 33.5135571710796 - type: nauc_ndcg_at_100_max value: 38.255675537823564 - type: nauc_ndcg_at_100_std value: -18.69039354080076 - type: nauc_ndcg_at_10_diff1 value: 33.04700239507369 - type: nauc_ndcg_at_10_max value: 38.87644726684572 - type: nauc_ndcg_at_10_std value: -21.761270791633518 - type: nauc_ndcg_at_1_diff1 value: 37.68273289251851 - type: nauc_ndcg_at_1_max value: 28.375793374298752 - type: nauc_ndcg_at_1_std value: -20.548630760150132 - type: nauc_ndcg_at_20_diff1 value: 33.59333929099863 - type: nauc_ndcg_at_20_max value: 39.13869119152796 - type: nauc_ndcg_at_20_std value: -20.455820914752028 - type: nauc_ndcg_at_3_diff1 value: 33.72195690786571 - type: nauc_ndcg_at_3_max value: 34.58224856532535 - type: nauc_ndcg_at_3_std value: -22.932493269664924 - type: nauc_ndcg_at_5_diff1 value: 33.454322211125806 - type: nauc_ndcg_at_5_max value: 37.62697388354373 - type: nauc_ndcg_at_5_std value: -22.471519101664132 - type: nauc_precision_at_1000_diff1 value: -4.815785976068792 - type: nauc_precision_at_1000_max value: -1.6093873845854942 - type: nauc_precision_at_1000_std value: 14.781030440554144 - type: nauc_precision_at_100_diff1 value: 11.770023400226492 - type: nauc_precision_at_100_max value: 32.39585905434347 - type: nauc_precision_at_100_std value: 13.926995268735807 - type: nauc_precision_at_10_diff1 value: 26.033870063028758 - type: nauc_precision_at_10_max value: 46.706875249128515 - type: nauc_precision_at_10_std value: -19.221899044995098 - type: nauc_precision_at_1_diff1 value: 37.68273289251851 - type: nauc_precision_at_1_max value: 28.375793374298752 - type: nauc_precision_at_1_std value: -20.548630760150132 - type: nauc_precision_at_20_diff1 value: 25.100441174579007 - type: nauc_precision_at_20_max value: 46.91020875403547 - type: nauc_precision_at_20_std value: -11.192277515531218 - type: nauc_precision_at_3_diff1 value: 30.618588495438985 - type: nauc_precision_at_3_max value: 37.248088037331286 - type: nauc_precision_at_3_std value: -23.92085440457614 - type: nauc_precision_at_5_diff1 value: 29.142344221391838 - type: nauc_precision_at_5_max value: 43.527892902769175 - type: nauc_precision_at_5_std value: -22.312841501376514 - type: nauc_recall_at_1000_diff1 value: 12.994211769214695 - type: nauc_recall_at_1000_max value: 55.743471097359446 - type: nauc_recall_at_1000_std value: 50.500646267896954 - type: nauc_recall_at_100_diff1 value: 25.84611790014738 - type: nauc_recall_at_100_max value: 62.84236269533729 - type: nauc_recall_at_100_std value: 16.99467383693571 - type: nauc_recall_at_10_diff1 value: 28.494332014682527 - type: nauc_recall_at_10_max value: 50.75293572531052 - type: nauc_recall_at_10_std value: -20.592936248452297 - type: nauc_recall_at_1_diff1 value: 37.89554066876773 - type: nauc_recall_at_1_max value: 28.579792597905413 - type: nauc_recall_at_1_std value: -20.51606339206856 - type: nauc_recall_at_20_diff1 value: 30.144206368539777 - type: nauc_recall_at_20_max value: 55.78415931465269 - type: nauc_recall_at_20_std value: -13.536686353112964 - type: nauc_recall_at_3_diff1 value: 31.153704257566993 - type: nauc_recall_at_3_max value: 38.10114875174283 - type: nauc_recall_at_3_std value: -24.098427776224725 - type: nauc_recall_at_5_diff1 value: 30.330462760076372 - type: nauc_recall_at_5_max value: 45.334521843132926 - type: nauc_recall_at_5_std value: -23.00539480314331 - type: ndcg_at_1 value: 23.281 - type: ndcg_at_10 value: 42.545 - type: ndcg_at_100 value: 48.091 - type: ndcg_at_1000 value: 49.135 - type: ndcg_at_20 value: 45.279 - type: ndcg_at_3 value: 34.507 - type: ndcg_at_5 value: 38.824 - type: precision_at_1 value: 23.281 - type: precision_at_10 value: 6.7250000000000005 - type: precision_at_100 value: 0.947 - type: precision_at_1000 value: 0.104 - type: precision_at_20 value: 3.9309999999999996 - type: precision_at_3 value: 14.771 - type: precision_at_5 value: 11.049000000000001 - type: recall_at_1 value: 22.655 - type: recall_at_10 value: 64.316 - type: recall_at_100 value: 89.596 - type: recall_at_1000 value: 97.627 - type: recall_at_20 value: 74.946 - type: recall_at_3 value: 42.625 - type: recall_at_5 value: 52.967 - task: type: Classification dataset: name: MTEB MTOPDomainClassification (en) type: mteb/mtop_domain config: en split: test revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf metrics: - type: accuracy value: 97.02462380300956 - type: f1 value: 96.7276440209508 - type: f1_weighted value: 97.04875399973407 - type: main_score value: 97.02462380300956 - task: type: Classification dataset: name: MTEB MTOPIntentClassification (en) type: mteb/mtop_intent config: en split: test revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba metrics: - type: accuracy value: 87.9252165982672 - type: f1 value: 67.80472291570956 - type: f1_weighted value: 87.85202683538105 - type: main_score value: 87.9252165982672 - task: type: Classification dataset: name: MTEB MassiveIntentClassification (en) type: mteb/amazon_massive_intent config: en split: test revision: 4672e20407010da34463acc759c162ca9734bca6 metrics: - type: accuracy value: 80.60524546065905 - type: f1 value: 78.33960315662881 - type: f1_weighted value: 79.95922500362934 - type: main_score value: 80.60524546065905 - task: type: Classification dataset: name: MTEB MassiveScenarioClassification (en) type: mteb/amazon_massive_scenario config: en split: test revision: fad2c6e8459f9e1c45d9315f4953d921437d70f8 metrics: - type: accuracy value: 81.93006052454606 - type: f1 value: 81.2714057686431 - type: f1_weighted value: 81.85548518216183 - type: main_score value: 81.93006052454606 - task: type: Clustering dataset: name: MTEB MedrxivClusteringP2P (default) type: mteb/medrxiv-clustering-p2p config: default split: test revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73 metrics: - type: main_score value: 41.53648349744648 - type: v_measure value: 41.53648349744648 - type: v_measure_std value: 1.3073220375465617 - task: type: Clustering dataset: name: MTEB MedrxivClusteringS2S (default) type: mteb/medrxiv-clustering-s2s config: default split: test revision: 35191c8c0dca72d8ff3efcd72aa802307d469663 metrics: - type: main_score value: 40.53587011806646 - type: v_measure value: 40.53587011806646 - type: v_measure_std value: 1.4167198988428324 - task: type: Reranking dataset: name: MTEB MindSmallReranking (default) type: mteb/mind_small config: default split: test revision: 59042f120c80e8afa9cdbb224f67076cec0fc9a7 metrics: - type: main_score value: 34.12179940649658 - type: map value: 34.12179940649658 - type: mrr value: 35.58575183247432 - type: nAUC_map_diff1 value: 18.455263729874243 - type: nAUC_map_max value: -18.69448732764168 - type: nAUC_map_std value: 8.198608386567457 - type: nAUC_mrr_diff1 value: 16.22907129322154 - type: nAUC_mrr_max value: -13.594180628663738 - type: nAUC_mrr_std value: 8.300689743851711 - task: type: Retrieval dataset: name: MTEB NFCorpus (default) type: mteb/nfcorpus config: default split: test revision: ec0fa4fe99da2ff19ca1214b7966684033a58814 metrics: - type: main_score value: 37.545 - type: map_at_1 value: 5.949999999999999 - type: map_at_10 value: 13.8 - type: map_at_100 value: 17.653 - type: map_at_1000 value: 19.322 - type: map_at_20 value: 15.318999999999999 - type: map_at_3 value: 10.211 - type: map_at_5 value: 11.757 - type: mrr_at_1 value: 49.84520123839009 - type: mrr_at_10 value: 58.26490245220894 - type: mrr_at_100 value: 58.751461262818694 - type: mrr_at_1000 value: 58.782721242595436 - type: mrr_at_20 value: 58.51537179710553 - type: mrr_at_3 value: 56.24355005159959 - type: mrr_at_5 value: 57.26522187822497 - type: nauc_map_at_1000_diff1 value: 24.708811804064975 - type: nauc_map_at_1000_max value: 22.51050994461675 - type: nauc_map_at_1000_std value: 12.29167974269923 - type: nauc_map_at_100_diff1 value: 27.740081813309736 - type: nauc_map_at_100_max value: 22.220395977232094 - type: nauc_map_at_100_std value: 8.670978243811184 - type: nauc_map_at_10_diff1 value: 34.4703308010072 - type: nauc_map_at_10_max value: 18.539226897919768 - type: nauc_map_at_10_std value: -0.9186030178287692 - type: nauc_map_at_1_diff1 value: 48.903950722167245 - type: nauc_map_at_1_max value: 4.368015121190565 - type: nauc_map_at_1_std value: -11.682230965520118 - type: nauc_map_at_20_diff1 value: 30.85072911235718 - type: nauc_map_at_20_max value: 20.024421045580016 - type: nauc_map_at_20_std value: 2.4437812527877223 - type: nauc_map_at_3_diff1 value: 39.1701521223124 - type: nauc_map_at_3_max value: 12.315298159822717 - type: nauc_map_at_3_std value: -5.1211175310668775 - type: nauc_map_at_5_diff1 value: 38.23279649034153 - type: nauc_map_at_5_max value: 14.562453378970972 - type: nauc_map_at_5_std value: -3.8872952078037306 - type: nauc_mrr_at_1000_diff1 value: 25.76454031603339 - type: nauc_mrr_at_1000_max value: 36.987486973646504 - type: nauc_mrr_at_1000_std value: 23.993127405911782 - type: nauc_mrr_at_100_diff1 value: 25.75748809964789 - type: nauc_mrr_at_100_max value: 37.00137109451698 - type: nauc_mrr_at_100_std value: 24.02115415632134 - type: nauc_mrr_at_10_diff1 value: 25.859969609083706 - type: nauc_mrr_at_10_max value: 36.94417043125623 - type: nauc_mrr_at_10_std value: 23.69193588816108 - type: nauc_mrr_at_1_diff1 value: 25.13856497503111 - type: nauc_mrr_at_1_max value: 33.3647833822104 - type: nauc_mrr_at_1_std value: 21.516825179743293 - type: nauc_mrr_at_20_diff1 value: 25.642602521543896 - type: nauc_mrr_at_20_max value: 37.00173585685738 - type: nauc_mrr_at_20_std value: 23.948759826317996 - type: nauc_mrr_at_3_diff1 value: 24.57379470383737 - type: nauc_mrr_at_3_max value: 35.05292254453142 - type: nauc_mrr_at_3_std value: 22.164140056553332 - type: nauc_mrr_at_5_diff1 value: 25.945828840033407 - type: nauc_mrr_at_5_max value: 36.28692013847132 - type: nauc_mrr_at_5_std value: 23.029834512220173 - type: nauc_ndcg_at_1000_diff1 value: 20.296757719323153 - type: nauc_ndcg_at_1000_max value: 37.48395095000081 - type: nauc_ndcg_at_1000_std value: 31.427363488785897 - type: nauc_ndcg_at_100_diff1 value: 20.850922448339382 - type: nauc_ndcg_at_100_max value: 31.994561388810332 - type: nauc_ndcg_at_100_std value: 24.999776113877374 - type: nauc_ndcg_at_10_diff1 value: 15.294338982345188 - type: nauc_ndcg_at_10_max value: 28.88313313311664 - type: nauc_ndcg_at_10_std value: 20.868634992089486 - type: nauc_ndcg_at_1_diff1 value: 26.184542545764266 - type: nauc_ndcg_at_1_max value: 33.49408854189648 - type: nauc_ndcg_at_1_std value: 21.644457229854616 - type: nauc_ndcg_at_20_diff1 value: 15.341797014632963 - type: nauc_ndcg_at_20_max value: 27.84956487113421 - type: nauc_ndcg_at_20_std value: 21.97010876262456 - type: nauc_ndcg_at_3_diff1 value: 16.617546176572887 - type: nauc_ndcg_at_3_max value: 31.0807079505372 - type: nauc_ndcg_at_3_std value: 20.563003372087447 - type: nauc_ndcg_at_5_diff1 value: 17.141262698147518 - type: nauc_ndcg_at_5_max value: 31.014000002769315 - type: nauc_ndcg_at_5_std value: 21.903989918122914 - type: nauc_precision_at_1000_diff1 value: -26.736915033118148 - type: nauc_precision_at_1000_max value: 0.41514162563304957 - type: nauc_precision_at_1000_std value: 29.414979920206335 - type: nauc_precision_at_100_diff1 value: -22.29841081134693 - type: nauc_precision_at_100_max value: 10.670850649163286 - type: nauc_precision_at_100_std value: 37.030209783550625 - type: nauc_precision_at_10_diff1 value: -7.401740939284052 - type: nauc_precision_at_10_max value: 26.372442015476512 - type: nauc_precision_at_10_std value: 28.058522245561985 - type: nauc_precision_at_1_diff1 value: 25.992836361025546 - type: nauc_precision_at_1_max value: 33.81712388873076 - type: nauc_precision_at_1_std value: 22.130100241561536 - type: nauc_precision_at_20_diff1 value: -14.715825716659179 - type: nauc_precision_at_20_max value: 21.0400050444382 - type: nauc_precision_at_20_std value: 32.37056700564148 - type: nauc_precision_at_3_diff1 value: 5.626852329412606 - type: nauc_precision_at_3_max value: 31.486758990096703 - type: nauc_precision_at_3_std value: 23.372250462239542 - type: nauc_precision_at_5_diff1 value: 1.2273456188651337 - type: nauc_precision_at_5_max value: 30.63448937975829 - type: nauc_precision_at_5_std value: 27.319392615570614 - type: nauc_recall_at_1000_diff1 value: 7.442058577828199 - type: nauc_recall_at_1000_max value: 17.366286208134948 - type: nauc_recall_at_1000_std value: 16.538023469059937 - type: nauc_recall_at_100_diff1 value: 18.263940318090828 - type: nauc_recall_at_100_max value: 18.766819889035368 - type: nauc_recall_at_100_std value: 10.297431485613268 - type: nauc_recall_at_10_diff1 value: 30.052808504776717 - type: nauc_recall_at_10_max value: 17.223636924464284 - type: nauc_recall_at_10_std value: -2.8915719805312126 - type: nauc_recall_at_1_diff1 value: 48.903950722167245 - type: nauc_recall_at_1_max value: 4.368015121190565 - type: nauc_recall_at_1_std value: -11.682230965520118 - type: nauc_recall_at_20_diff1 value: 25.00678345922952 - type: nauc_recall_at_20_max value: 17.734815525743993 - type: nauc_recall_at_20_std value: 1.2937788396283523 - type: nauc_recall_at_3_diff1 value: 34.053479666933164 - type: nauc_recall_at_3_max value: 10.356061180744728 - type: nauc_recall_at_3_std value: -7.622782189103819 - type: nauc_recall_at_5_diff1 value: 35.282050319114994 - type: nauc_recall_at_5_max value: 13.444414495259005 - type: nauc_recall_at_5_std value: -6.406123122708332 - type: ndcg_at_1 value: 47.833 - type: ndcg_at_10 value: 37.545 - type: ndcg_at_100 value: 34.608 - type: ndcg_at_1000 value: 43.789 - type: ndcg_at_20 value: 34.724 - type: ndcg_at_3 value: 43.055 - type: ndcg_at_5 value: 40.595 - type: precision_at_1 value: 49.536 - type: precision_at_10 value: 27.678000000000004 - type: precision_at_100 value: 8.901 - type: precision_at_1000 value: 2.225 - type: precision_at_20 value: 20.279 - type: precision_at_3 value: 40.144000000000005 - type: precision_at_5 value: 34.675 - type: recall_at_1 value: 5.949999999999999 - type: recall_at_10 value: 18.368000000000002 - type: recall_at_100 value: 36.702 - type: recall_at_1000 value: 69.39800000000001 - type: recall_at_20 value: 22.241 - type: recall_at_3 value: 11.618 - type: recall_at_5 value: 14.338999999999999 - task: type: Retrieval dataset: name: MTEB NQ (default) type: mteb/nq config: default split: test revision: b774495ed302d8c44a3a7ea25c90dbce03968f31 metrics: - type: main_score value: 64.693 - type: map_at_1 value: 40.119 - type: map_at_10 value: 57.008 - type: map_at_100 value: 57.769999999999996 - type: map_at_1000 value: 57.782999999999994 - type: map_at_20 value: 57.528999999999996 - type: map_at_3 value: 52.713 - type: map_at_5 value: 55.462 - type: mrr_at_1 value: 45.017381228273464 - type: mrr_at_10 value: 59.62700481892242 - type: mrr_at_100 value: 60.11977007964554 - type: mrr_at_1000 value: 60.12838314206039 - type: mrr_at_20 value: 59.96971543639854 - type: mrr_at_3 value: 56.38277327153322 - type: mrr_at_5 value: 58.559772112784756 - type: nauc_map_at_1000_diff1 value: 39.016224863361316 - type: nauc_map_at_1000_max value: 30.677526741613914 - type: nauc_map_at_1000_std value: -9.925306326190029 - type: nauc_map_at_100_diff1 value: 39.02038091276591 - type: nauc_map_at_100_max value: 30.687899774856326 - type: nauc_map_at_100_std value: -9.914518833390677 - type: nauc_map_at_10_diff1 value: 39.04523753783543 - type: nauc_map_at_10_max value: 30.976052448225627 - type: nauc_map_at_10_std value: -10.41607954987974 - type: nauc_map_at_1_diff1 value: 40.06219774868448 - type: nauc_map_at_1_max value: 26.735486652072517 - type: nauc_map_at_1_std value: -8.304382193524896 - type: nauc_map_at_20_diff1 value: 39.05577477358533 - type: nauc_map_at_20_max value: 30.78179300885049 - type: nauc_map_at_20_std value: -10.033002471334074 - type: nauc_map_at_3_diff1 value: 38.802559695913885 - type: nauc_map_at_3_max value: 30.365699555342978 - type: nauc_map_at_3_std value: -11.716942405728881 - type: nauc_map_at_5_diff1 value: 38.88593641854277 - type: nauc_map_at_5_max value: 30.93585211223555 - type: nauc_map_at_5_std value: -10.926633622752911 - type: nauc_mrr_at_1000_diff1 value: 39.04692080086692 - type: nauc_mrr_at_1000_max value: 30.197468259524175 - type: nauc_mrr_at_1000_std value: -7.818491692833017 - type: nauc_mrr_at_100_diff1 value: 39.0473015118493 - type: nauc_mrr_at_100_max value: 30.203218891973965 - type: nauc_mrr_at_100_std value: -7.809410627895269 - type: nauc_mrr_at_10_diff1 value: 39.022526617566456 - type: nauc_mrr_at_10_max value: 30.41103199763037 - type: nauc_mrr_at_10_std value: -7.986473780645788 - type: nauc_mrr_at_1_diff1 value: 40.2687402313342 - type: nauc_mrr_at_1_max value: 26.56359606867155 - type: nauc_mrr_at_1_std value: -6.6659359448538025 - type: nauc_mrr_at_20_diff1 value: 39.048111884686826 - type: nauc_mrr_at_20_max value: 30.246914959156364 - type: nauc_mrr_at_20_std value: -7.801804075454251 - type: nauc_mrr_at_3_diff1 value: 38.8647060004973 - type: nauc_mrr_at_3_max value: 30.225427021287963 - type: nauc_mrr_at_3_std value: -9.016676247800575 - type: nauc_mrr_at_5_diff1 value: 38.95589884289447 - type: nauc_mrr_at_5_max value: 30.55482027762662 - type: nauc_mrr_at_5_std value: -8.287991164740555 - type: nauc_ndcg_at_1000_diff1 value: 38.935229352725536 - type: nauc_ndcg_at_1000_max value: 31.318278701790277 - type: nauc_ndcg_at_1000_std value: -8.498883716013742 - type: nauc_ndcg_at_100_diff1 value: 39.00131687376651 - type: nauc_ndcg_at_100_max value: 31.60126452179523 - type: nauc_ndcg_at_100_std value: -8.0878761098937 - type: nauc_ndcg_at_10_diff1 value: 38.997637272745976 - type: nauc_ndcg_at_10_max value: 32.81562205617119 - type: nauc_ndcg_at_10_std value: -9.809117549403716 - type: nauc_ndcg_at_1_diff1 value: 40.2687402313342 - type: nauc_ndcg_at_1_max value: 26.56359606867155 - type: nauc_ndcg_at_1_std value: -6.6659359448538025 - type: nauc_ndcg_at_20_diff1 value: 39.05787809282005 - type: nauc_ndcg_at_20_max value: 32.148837127474216 - type: nauc_ndcg_at_20_std value: -8.538216720226362 - type: nauc_ndcg_at_3_diff1 value: 38.514904225460185 - type: nauc_ndcg_at_3_max value: 31.647932572190907 - type: nauc_ndcg_at_3_std value: -12.117323520301271 - type: nauc_ndcg_at_5_diff1 value: 38.67523620158631 - type: nauc_ndcg_at_5_max value: 32.71111428248374 - type: nauc_ndcg_at_5_std value: -10.830509911489106 - type: nauc_precision_at_1000_diff1 value: -10.134425320872637 - type: nauc_precision_at_1000_max value: -7.9214866985442836 - type: nauc_precision_at_1000_std value: 14.593125138517463 - type: nauc_precision_at_100_diff1 value: -6.427184925035445 - type: nauc_precision_at_100_max value: -3.565171885582329 - type: nauc_precision_at_100_std value: 15.87161403279646 - type: nauc_precision_at_10_diff1 value: 9.87974963974257 - type: nauc_precision_at_10_max value: 14.701681974930208 - type: nauc_precision_at_10_std value: 3.7336847482921924 - type: nauc_precision_at_1_diff1 value: 40.2687402313342 - type: nauc_precision_at_1_max value: 26.56359606867155 - type: nauc_precision_at_1_std value: -6.6659359448538025 - type: nauc_precision_at_20_diff1 value: 2.890969722929749 - type: nauc_precision_at_20_max value: 6.794303444012595 - type: nauc_precision_at_20_std value: 10.705845010583102 - type: nauc_precision_at_3_diff1 value: 25.828531407512916 - type: nauc_precision_at_3_max value: 26.003194922700068 - type: nauc_precision_at_3_std value: -9.365843001852745 - type: nauc_precision_at_5_diff1 value: 18.442430286590213 - type: nauc_precision_at_5_max value: 22.17126455978319 - type: nauc_precision_at_5_std value: -3.307912326207094 - type: nauc_recall_at_1000_diff1 value: 37.08230039820157 - type: nauc_recall_at_1000_max value: 47.10529218716289 - type: nauc_recall_at_1000_std value: 47.786964110589096 - type: nauc_recall_at_100_diff1 value: 41.32053677940413 - type: nauc_recall_at_100_max value: 53.09289155866568 - type: nauc_recall_at_100_std value: 32.47492854799267 - type: nauc_recall_at_10_diff1 value: 37.31427344851398 - type: nauc_recall_at_10_max value: 43.0702780767873 - type: nauc_recall_at_10_std value: -10.887409444200305 - type: nauc_recall_at_1_diff1 value: 40.06219774868448 - type: nauc_recall_at_1_max value: 26.735486652072517 - type: nauc_recall_at_1_std value: -8.304382193524896 - type: nauc_recall_at_20_diff1 value: 38.026247692487225 - type: nauc_recall_at_20_max value: 43.122612480943125 - type: nauc_recall_at_20_std value: 0.06425536869830446 - type: nauc_recall_at_3_diff1 value: 36.42120384763962 - type: nauc_recall_at_3_max value: 34.94129978903372 - type: nauc_recall_at_3_std value: -15.716640140198779 - type: nauc_recall_at_5_diff1 value: 36.15895636103322 - type: nauc_recall_at_5_max value: 38.80623578799298 - type: nauc_recall_at_5_std value: -13.51525373978869 - type: ndcg_at_1 value: 45.017 - type: ndcg_at_10 value: 64.693 - type: ndcg_at_100 value: 67.632 - type: ndcg_at_1000 value: 67.91199999999999 - type: ndcg_at_20 value: 66.277 - type: ndcg_at_3 value: 57.046 - type: ndcg_at_5 value: 61.516999999999996 - type: precision_at_1 value: 45.017 - type: precision_at_10 value: 10.18 - type: precision_at_100 value: 1.185 - type: precision_at_1000 value: 0.121 - type: precision_at_20 value: 5.479 - type: precision_at_3 value: 25.541000000000004 - type: precision_at_5 value: 17.949 - type: recall_at_1 value: 40.119 - type: recall_at_10 value: 85.139 - type: recall_at_100 value: 97.444 - type: recall_at_1000 value: 99.529 - type: recall_at_20 value: 90.88199999999999 - type: recall_at_3 value: 65.88000000000001 - type: recall_at_5 value: 76.132 - task: type: Retrieval dataset: name: MTEB QuoraRetrieval (default) type: mteb/quora config: default split: test revision: e4e08e0b7dbe3c8700f0daef558ff32256715259 metrics: - type: main_score value: 88.773 - type: map_at_1 value: 70.96000000000001 - type: map_at_10 value: 85.174 - type: map_at_100 value: 85.804 - type: map_at_1000 value: 85.817 - type: map_at_20 value: 85.596 - type: map_at_3 value: 82.219 - type: map_at_5 value: 84.098 - type: mrr_at_1 value: 81.76 - type: mrr_at_10 value: 87.79770634920607 - type: mrr_at_100 value: 87.89799102352673 - type: mrr_at_1000 value: 87.89865476743903 - type: mrr_at_20 value: 87.87680512328197 - type: mrr_at_3 value: 86.81999999999978 - type: mrr_at_5 value: 87.51299999999969 - type: nauc_map_at_1000_diff1 value: 76.90119675123604 - type: nauc_map_at_1000_max value: 20.079761155170527 - type: nauc_map_at_1000_std value: -62.08844878736319 - type: nauc_map_at_100_diff1 value: 76.91315659037733 - type: nauc_map_at_100_max value: 20.037613519830543 - type: nauc_map_at_100_std value: -62.1809605413574 - type: nauc_map_at_10_diff1 value: 77.29490073584684 - type: nauc_map_at_10_max value: 18.97493585375514 - type: nauc_map_at_10_std value: -65.06133578431042 - type: nauc_map_at_1_diff1 value: 80.92204517914038 - type: nauc_map_at_1_max value: 12.955779715044127 - type: nauc_map_at_1_std value: -53.185870692847814 - type: nauc_map_at_20_diff1 value: 77.06595372320452 - type: nauc_map_at_20_max value: 19.587544100405307 - type: nauc_map_at_20_std value: -63.38932039031718 - type: nauc_map_at_3_diff1 value: 77.81358593606132 - type: nauc_map_at_3_max value: 16.415667501797888 - type: nauc_map_at_3_std value: -65.91817124009025 - type: nauc_map_at_5_diff1 value: 77.55572041802866 - type: nauc_map_at_5_max value: 17.84810271472641 - type: nauc_map_at_5_std value: -66.5202429218229 - type: nauc_mrr_at_1000_diff1 value: 77.25919152483527 - type: nauc_mrr_at_1000_max value: 23.266505681060313 - type: nauc_mrr_at_1000_std value: -56.997207262592106 - type: nauc_mrr_at_100_diff1 value: 77.25865200926027 - type: nauc_mrr_at_100_max value: 23.266917952901537 - type: nauc_mrr_at_100_std value: -56.99775622461676 - type: nauc_mrr_at_10_diff1 value: 77.27177237809222 - type: nauc_mrr_at_10_max value: 23.234422413279194 - type: nauc_mrr_at_10_std value: -57.287098821203166 - type: nauc_mrr_at_1_diff1 value: 77.87705968629228 - type: nauc_mrr_at_1_max value: 23.357364820166353 - type: nauc_mrr_at_1_std value: -52.724677718394254 - type: nauc_mrr_at_20_diff1 value: 77.26510245027495 - type: nauc_mrr_at_20_max value: 23.250601444229872 - type: nauc_mrr_at_20_std value: -57.073576665896155 - type: nauc_mrr_at_3_diff1 value: 77.08835110871802 - type: nauc_mrr_at_3_max value: 23.37973990414157 - type: nauc_mrr_at_3_std value: -57.54668286148783 - type: nauc_mrr_at_5_diff1 value: 77.22940631493309 - type: nauc_mrr_at_5_max value: 23.256197542861436 - type: nauc_mrr_at_5_std value: -57.710428425249404 - type: nauc_ndcg_at_1000_diff1 value: 76.67905982606639 - type: nauc_ndcg_at_1000_max value: 21.96802435224643 - type: nauc_ndcg_at_1000_std value: -59.660695538408405 - type: nauc_ndcg_at_100_diff1 value: 76.72641745578917 - type: nauc_ndcg_at_100_max value: 21.752538833557992 - type: nauc_ndcg_at_100_std value: -60.14387533073589 - type: nauc_ndcg_at_10_diff1 value: 77.1697583832975 - type: nauc_ndcg_at_10_max value: 19.90438134636175 - type: nauc_ndcg_at_10_std value: -65.62207352990609 - type: nauc_ndcg_at_1_diff1 value: 77.87705968629228 - type: nauc_ndcg_at_1_max value: 23.357364820166353 - type: nauc_ndcg_at_1_std value: -52.724677718394254 - type: nauc_ndcg_at_20_diff1 value: 76.98061052184806 - type: nauc_ndcg_at_20_max value: 20.514885434747328 - type: nauc_ndcg_at_20_std value: -63.237149791291415 - type: nauc_ndcg_at_3_diff1 value: 76.32552624964065 - type: nauc_ndcg_at_3_max value: 19.923840442393544 - type: nauc_ndcg_at_3_std value: -63.588842129524245 - type: nauc_ndcg_at_5_diff1 value: 76.9533163521833 - type: nauc_ndcg_at_5_max value: 19.51602820692585 - type: nauc_ndcg_at_5_std value: -66.23316232094454 - type: nauc_precision_at_1000_diff1 value: -45.73706664910733 - type: nauc_precision_at_1000_max value: 7.913436156563994 - type: nauc_precision_at_1000_std value: 53.06948338411226 - type: nauc_precision_at_100_diff1 value: -45.31368947263091 - type: nauc_precision_at_100_max value: 7.188900869218029 - type: nauc_precision_at_100_std value: 50.86284056359611 - type: nauc_precision_at_10_diff1 value: -39.50336694936736 - type: nauc_precision_at_10_max value: 6.702378324096768 - type: nauc_precision_at_10_std value: 31.03637763595204 - type: nauc_precision_at_1_diff1 value: 77.87705968629228 - type: nauc_precision_at_1_max value: 23.357364820166353 - type: nauc_precision_at_1_std value: -52.724677718394254 - type: nauc_precision_at_20_diff1 value: -43.09729408672091 - type: nauc_precision_at_20_max value: 6.532907159014953 - type: nauc_precision_at_20_std value: 40.98770041852758 - type: nauc_precision_at_3_diff1 value: -19.675745078316503 - type: nauc_precision_at_3_max value: 9.254372245883973 - type: nauc_precision_at_3_std value: 3.557752877438361 - type: nauc_precision_at_5_diff1 value: -32.17451238619065 - type: nauc_precision_at_5_max value: 7.457382998315637 - type: nauc_precision_at_5_std value: 17.684523480181884 - type: nauc_recall_at_1000_diff1 value: 35.54833030189762 - type: nauc_recall_at_1000_max value: -113.13072963237435 - type: nauc_recall_at_1000_std value: -45.37230224613866 - type: nauc_recall_at_100_diff1 value: 74.70783770156788 - type: nauc_recall_at_100_max value: 5.165483155761366 - type: nauc_recall_at_100_std value: -98.18356589742223 - type: nauc_recall_at_10_diff1 value: 76.44831137766471 - type: nauc_recall_at_10_max value: 6.645874880559598 - type: nauc_recall_at_10_std value: -104.42733750490795 - type: nauc_recall_at_1_diff1 value: 80.92204517914038 - type: nauc_recall_at_1_max value: 12.955779715044127 - type: nauc_recall_at_1_std value: -53.185870692847814 - type: nauc_recall_at_20_diff1 value: 76.9330017100496 - type: nauc_recall_at_20_max value: 1.3282965733900722 - type: nauc_recall_at_20_std value: -110.44267520170585 - type: nauc_recall_at_3_diff1 value: 74.75571112449231 - type: nauc_recall_at_3_max value: 11.392712834655518 - type: nauc_recall_at_3_std value: -77.70319541112546 - type: nauc_recall_at_5_diff1 value: 74.44393885573719 - type: nauc_recall_at_5_max value: 9.071230160466847 - type: nauc_recall_at_5_std value: -90.6015799064062 - type: ndcg_at_1 value: 81.76 - type: ndcg_at_10 value: 88.773 - type: ndcg_at_100 value: 89.93100000000001 - type: ndcg_at_1000 value: 90.005 - type: ndcg_at_20 value: 89.436 - type: ndcg_at_3 value: 85.997 - type: ndcg_at_5 value: 87.571 - type: precision_at_1 value: 81.76 - type: precision_at_10 value: 13.542000000000002 - type: precision_at_100 value: 1.538 - type: precision_at_1000 value: 0.157 - type: precision_at_20 value: 7.184 - type: precision_at_3 value: 37.8 - type: precision_at_5 value: 24.898 - type: recall_at_1 value: 70.96000000000001 - type: recall_at_10 value: 95.741 - type: recall_at_100 value: 99.685 - type: recall_at_1000 value: 99.995 - type: recall_at_20 value: 97.909 - type: recall_at_3 value: 87.739 - type: recall_at_5 value: 92.203 - task: type: Clustering dataset: name: MTEB RedditClustering (default) type: mteb/reddit-clustering config: default split: test revision: 24640382cdbf8abc73003fb0fa6d111a705499eb metrics: - type: main_score value: 65.91810902432418 - type: v_measure value: 65.91810902432418 - type: v_measure_std value: 3.988775454635202 - task: type: Clustering dataset: name: MTEB RedditClusteringP2P (default) type: mteb/reddit-clustering-p2p config: default split: test revision: 385e3cb46b4cfa89021f56c4380204149d0efe33 metrics: - type: main_score value: 68.02321609158898 - type: v_measure value: 68.02321609158898 - type: v_measure_std value: 13.048787017567099 - task: type: Retrieval dataset: name: MTEB SCIDOCS (default) type: mteb/scidocs config: default split: test revision: f8c2fcf00f625baaa80f62ec5bd9e1fff3b8ae88 metrics: - type: main_score value: 23.814 - type: map_at_1 value: 5.455 - type: map_at_10 value: 14.208000000000002 - type: map_at_100 value: 17.328 - type: map_at_1000 value: 17.748 - type: map_at_20 value: 15.735 - type: map_at_3 value: 9.614 - type: map_at_5 value: 11.777999999999999 - type: mrr_at_1 value: 26.900000000000002 - type: mrr_at_10 value: 38.7683333333333 - type: mrr_at_100 value: 39.86887355087612 - type: mrr_at_1000 value: 39.89416581484622 - type: mrr_at_20 value: 39.45076687545336 - type: mrr_at_3 value: 34.283333333333324 - type: mrr_at_5 value: 36.7733333333333 - type: nauc_map_at_1000_diff1 value: 6.582032797360866 - type: nauc_map_at_1000_max value: 17.29971642208067 - type: nauc_map_at_1000_std value: 6.571653079053965 - type: nauc_map_at_100_diff1 value: 6.6520274055220945 - type: nauc_map_at_100_max value: 17.28927582943446 - type: nauc_map_at_100_std value: 6.3788070086997735 - type: nauc_map_at_10_diff1 value: 7.8097868021789765 - type: nauc_map_at_10_max value: 15.868814598414307 - type: nauc_map_at_10_std value: 1.3833485160000003 - type: nauc_map_at_1_diff1 value: 20.002393048021077 - type: nauc_map_at_1_max value: 16.777673629413144 - type: nauc_map_at_1_std value: -1.5982142140773345 - type: nauc_map_at_20_diff1 value: 7.026484961291383 - type: nauc_map_at_20_max value: 16.358039615308098 - type: nauc_map_at_20_std value: 4.265555678748822 - type: nauc_map_at_3_diff1 value: 11.670235117521639 - type: nauc_map_at_3_max value: 15.421371305785032 - type: nauc_map_at_3_std value: -2.0891385987905253 - type: nauc_map_at_5_diff1 value: 8.782941099433515 - type: nauc_map_at_5_max value: 15.429505319062791 - type: nauc_map_at_5_std value: 0.01706038881959217 - type: nauc_mrr_at_1000_diff1 value: 14.424089575104654 - type: nauc_mrr_at_1000_max value: 18.354632635310146 - type: nauc_mrr_at_1000_std value: 3.148669746271006 - type: nauc_mrr_at_100_diff1 value: 14.43190469520255 - type: nauc_mrr_at_100_max value: 18.37445314994635 - type: nauc_mrr_at_100_std value: 3.175095104402658 - type: nauc_mrr_at_10_diff1 value: 14.015953357582356 - type: nauc_mrr_at_10_max value: 18.334773185007375 - type: nauc_mrr_at_10_std value: 3.1788218175601917 - type: nauc_mrr_at_1_diff1 value: 20.06438180516676 - type: nauc_mrr_at_1_max value: 16.906770193671957 - type: nauc_mrr_at_1_std value: -1.591329233808127 - type: nauc_mrr_at_20_diff1 value: 14.126339493553159 - type: nauc_mrr_at_20_max value: 18.316449447055653 - type: nauc_mrr_at_20_std value: 3.1850941428621042 - type: nauc_mrr_at_3_diff1 value: 14.730386161975737 - type: nauc_mrr_at_3_max value: 17.32498171231654 - type: nauc_mrr_at_3_std value: 1.321654906709584 - type: nauc_mrr_at_5_diff1 value: 14.46476336413886 - type: nauc_mrr_at_5_max value: 17.940958841978826 - type: nauc_mrr_at_5_std value: 2.9529508335708945 - type: nauc_ndcg_at_1000_diff1 value: 6.681346718194129 - type: nauc_ndcg_at_1000_max value: 21.404613477283746 - type: nauc_ndcg_at_1000_std value: 14.596655479547055 - type: nauc_ndcg_at_100_diff1 value: 6.3302594607492155 - type: nauc_ndcg_at_100_max value: 21.26459769654865 - type: nauc_ndcg_at_100_std value: 14.522962033467959 - type: nauc_ndcg_at_10_diff1 value: 7.025732359853311 - type: nauc_ndcg_at_10_max value: 17.31881906701822 - type: nauc_ndcg_at_10_std value: 4.692540938431521 - type: nauc_ndcg_at_1_diff1 value: 20.06438180516676 - type: nauc_ndcg_at_1_max value: 16.906770193671957 - type: nauc_ndcg_at_1_std value: -1.591329233808127 - type: nauc_ndcg_at_20_diff1 value: 6.355140893975436 - type: nauc_ndcg_at_20_max value: 18.29467935307024 - type: nauc_ndcg_at_20_std value: 8.87309764856374 - type: nauc_ndcg_at_3_diff1 value: 11.131091987737578 - type: nauc_ndcg_at_3_max value: 15.876946297140213 - type: nauc_ndcg_at_3_std value: -0.19961674229045062 - type: nauc_ndcg_at_5_diff1 value: 8.719384001108486 - type: nauc_ndcg_at_5_max value: 16.561854761839523 - type: nauc_ndcg_at_5_std value: 2.849455858958004 - type: nauc_precision_at_1000_diff1 value: -3.264266561841031 - type: nauc_precision_at_1000_max value: 27.054907731659355 - type: nauc_precision_at_1000_std value: 42.6582722652614 - type: nauc_precision_at_100_diff1 value: -1.4147583046219077 - type: nauc_precision_at_100_max value: 22.691769918104637 - type: nauc_precision_at_100_std value: 30.417860777083998 - type: nauc_precision_at_10_diff1 value: 0.7460714765387558 - type: nauc_precision_at_10_max value: 16.189155199570223 - type: nauc_precision_at_10_std value: 8.466856326540606 - type: nauc_precision_at_1_diff1 value: 20.06438180516676 - type: nauc_precision_at_1_max value: 16.906770193671957 - type: nauc_precision_at_1_std value: -1.591329233808127 - type: nauc_precision_at_20_diff1 value: -0.29107581757496714 - type: nauc_precision_at_20_max value: 17.13909220544385 - type: nauc_precision_at_20_std value: 16.413326815174717 - type: nauc_precision_at_3_diff1 value: 7.101179998696147 - type: nauc_precision_at_3_max value: 14.797248842818975 - type: nauc_precision_at_3_std value: 0.40582828085273265 - type: nauc_precision_at_5_diff1 value: 3.4483179666389696 - type: nauc_precision_at_5_max value: 15.735507259648934 - type: nauc_precision_at_5_std value: 5.671451893149887 - type: nauc_recall_at_1000_diff1 value: -3.8075718189695547 - type: nauc_recall_at_1000_max value: 27.218180153734124 - type: nauc_recall_at_1000_std value: 44.46679820329153 - type: nauc_recall_at_100_diff1 value: -1.4536649156519559 - type: nauc_recall_at_100_max value: 22.44690502045992 - type: nauc_recall_at_100_std value: 30.235557227945275 - type: nauc_recall_at_10_diff1 value: 0.6119379049099861 - type: nauc_recall_at_10_max value: 15.882135185205446 - type: nauc_recall_at_10_std value: 8.176733663905573 - type: nauc_recall_at_1_diff1 value: 20.002393048021077 - type: nauc_recall_at_1_max value: 16.777673629413144 - type: nauc_recall_at_1_std value: -1.5982142140773345 - type: nauc_recall_at_20_diff1 value: -0.1682800060016626 - type: nauc_recall_at_20_max value: 16.971491120013564 - type: nauc_recall_at_20_std value: 16.122046383351293 - type: nauc_recall_at_3_diff1 value: 6.988663029514718 - type: nauc_recall_at_3_max value: 14.528152900658856 - type: nauc_recall_at_3_std value: 0.17590933968510467 - type: nauc_recall_at_5_diff1 value: 3.353041984845736 - type: nauc_recall_at_5_max value: 15.403568054057326 - type: nauc_recall_at_5_std value: 5.319244399661828 - type: ndcg_at_1 value: 26.900000000000002 - type: ndcg_at_10 value: 23.814 - type: ndcg_at_100 value: 34.943999999999996 - type: ndcg_at_1000 value: 40.78 - type: ndcg_at_20 value: 27.643 - type: ndcg_at_3 value: 21.227 - type: ndcg_at_5 value: 19.038 - type: precision_at_1 value: 26.900000000000002 - type: precision_at_10 value: 12.73 - type: precision_at_100 value: 2.881 - type: precision_at_1000 value: 0.426 - type: precision_at_20 value: 8.57 - type: precision_at_3 value: 19.6 - type: precision_at_5 value: 16.8 - type: recall_at_1 value: 5.455 - type: recall_at_10 value: 25.802999999999997 - type: recall_at_100 value: 58.45 - type: recall_at_1000 value: 86.457 - type: recall_at_20 value: 34.762 - type: recall_at_3 value: 11.943 - type: recall_at_5 value: 17.043 - task: type: STS dataset: name: MTEB SICK-R (default) type: mteb/sickr-sts config: default split: test revision: 20a6d6f312dd54037fe07a32d58e5e168867909d metrics: - type: cosine_pearson value: 85.60157402941752 - type: cosine_spearman value: 82.98956725441452 - type: euclidean_pearson value: 83.07824357271161 - type: euclidean_spearman value: 82.98957395335212 - type: main_score value: 82.98956725441452 - type: manhattan_pearson value: 83.10748351148622 - type: manhattan_spearman value: 83.16217281563378 - type: pearson value: 85.60157402941752 - type: spearman value: 82.98956725441452 - task: type: STS dataset: name: MTEB STS12 (default) type: mteb/sts12-sts config: default split: test revision: a0d554a64d88156834ff5ae9920b964011b16384 metrics: - type: cosine_pearson value: 85.20198919395854 - type: cosine_spearman value: 78.17308450713497 - type: euclidean_pearson value: 82.91465813078975 - type: euclidean_spearman value: 78.17308450713497 - type: main_score value: 78.17308450713497 - type: manhattan_pearson value: 83.36938760055344 - type: manhattan_spearman value: 78.77166023561925 - type: pearson value: 85.20198919395854 - type: spearman value: 78.17308450713497 - task: type: STS dataset: name: MTEB STS13 (default) type: mteb/sts13-sts config: default split: test revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca metrics: - type: cosine_pearson value: 87.3197290035165 - type: cosine_spearman value: 88.12589189918039 - type: euclidean_pearson value: 87.88474436451652 - type: euclidean_spearman value: 88.12589189918039 - type: main_score value: 88.12589189918039 - type: manhattan_pearson value: 88.1114243109502 - type: manhattan_spearman value: 88.40111910955112 - type: pearson value: 87.3197290035165 - type: spearman value: 88.12589189918039 - task: type: STS dataset: name: MTEB STS15 (default) type: mteb/sts15-sts config: default split: test revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3 metrics: - type: cosine_pearson value: 87.91424745154934 - type: cosine_spearman value: 88.78510857775494 - type: euclidean_pearson value: 88.60854825357943 - type: euclidean_spearman value: 88.78511307332248 - type: main_score value: 88.78510857775494 - type: manhattan_pearson value: 88.81490531409946 - type: manhattan_spearman value: 89.10162579991359 - type: pearson value: 87.91424745154934 - type: spearman value: 88.78510857775494 - task: type: STS dataset: name: MTEB STS16 (default) type: mteb/sts16-sts config: default split: test revision: 4d8694f8f0e0100860b497b999b3dbed754a0513 metrics: - type: cosine_pearson value: 84.42255273136605 - type: cosine_spearman value: 86.46810322536955 - type: euclidean_pearson value: 86.255541184091 - type: euclidean_spearman value: 86.46810322536955 - type: main_score value: 86.46810322536955 - type: manhattan_pearson value: 86.72678851651064 - type: manhattan_spearman value: 86.93777990302539 - type: pearson value: 84.42255273136605 - type: spearman value: 86.46810322536955 - task: type: STS dataset: name: MTEB STS17 (en-en) type: mteb/sts17-crosslingual-sts config: en-en split: test revision: faeb762787bd10488a50c8b5be4a3b82e411949c metrics: - type: cosine_pearson value: 91.72746389892356 - type: cosine_spearman value: 92.23283881812245 - type: euclidean_pearson value: 92.29179177488737 - type: euclidean_spearman value: 92.23283881812245 - type: main_score value: 92.23283881812245 - type: manhattan_pearson value: 92.13764526009247 - type: manhattan_spearman value: 92.0582843442798 - type: pearson value: 91.72746389892356 - type: spearman value: 92.23283881812245 - task: type: STS dataset: name: MTEB STSBenchmark (default) type: mteb/stsbenchmark-sts config: default split: test revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831 metrics: - type: cosine_pearson value: 86.14912927994007 - type: cosine_spearman value: 87.46655844472012 - type: euclidean_pearson value: 87.53026653408118 - type: euclidean_spearman value: 87.46655844472012 - type: main_score value: 87.46655844472012 - type: manhattan_pearson value: 87.68289898403299 - type: manhattan_spearman value: 87.73630507998439 - type: pearson value: 86.14912927994007 - type: spearman value: 87.46655844472012 - task: type: Reranking dataset: name: MTEB SciDocsRR (default) type: mteb/scidocs-reranking config: default split: test revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab metrics: - type: main_score value: 86.97859154411299 - type: map value: 86.97859154411299 - type: mrr value: 96.35598968932302 - type: nAUC_map_diff1 value: -18.506120190268017 - type: nAUC_map_max value: 55.78442121746724 - type: nAUC_map_std value: 66.27889919160313 - type: nAUC_mrr_diff1 value: 18.288014199762895 - type: nAUC_mrr_max value: 83.25297655347828 - type: nAUC_mrr_std value: 72.809885375971 - task: type: Retrieval dataset: name: MTEB SciFact (default) type: mteb/scifact config: default split: test revision: 0228b52cf27578f30900b9e5271d331663a030d7 metrics: - type: main_score value: 52.842 - type: map_at_1 value: 32.911 - type: map_at_10 value: 46.013 - type: map_at_100 value: 47.11 - type: map_at_1000 value: 47.137 - type: map_at_20 value: 46.78 - type: map_at_3 value: 41.900999999999996 - type: map_at_5 value: 44.357 - type: mrr_at_1 value: 35.0 - type: mrr_at_10 value: 46.96574074074072 - type: mrr_at_100 value: 47.959931245967184 - type: mrr_at_1000 value: 47.98510849619688 - type: mrr_at_20 value: 47.68440206880607 - type: mrr_at_3 value: 43.77777777777776 - type: mrr_at_5 value: 45.611111111111086 - type: nauc_map_at_1000_diff1 value: 42.89180178126247 - type: nauc_map_at_1000_max value: 45.75105611403444 - type: nauc_map_at_1000_std value: 17.463513608950578 - type: nauc_map_at_100_diff1 value: 42.893512582653656 - type: nauc_map_at_100_max value: 45.754617699990035 - type: nauc_map_at_100_std value: 17.490513656867037 - type: nauc_map_at_10_diff1 value: 42.364748689290415 - type: nauc_map_at_10_max value: 45.56642444523947 - type: nauc_map_at_10_std value: 17.079579644716894 - type: nauc_map_at_1_diff1 value: 48.949793800671124 - type: nauc_map_at_1_max value: 45.82239538118238 - type: nauc_map_at_1_std value: 11.183927196674755 - type: nauc_map_at_20_diff1 value: 42.67947282270775 - type: nauc_map_at_20_max value: 45.62274524098362 - type: nauc_map_at_20_std value: 17.51316198529124 - type: nauc_map_at_3_diff1 value: 43.238404886755745 - type: nauc_map_at_3_max value: 43.350130089078895 - type: nauc_map_at_3_std value: 14.13657834477199 - type: nauc_map_at_5_diff1 value: 42.54474356788842 - type: nauc_map_at_5_max value: 44.75146781225222 - type: nauc_map_at_5_std value: 16.15648396925114 - type: nauc_mrr_at_1000_diff1 value: 43.556859926201554 - type: nauc_mrr_at_1000_max value: 47.140291020802906 - type: nauc_mrr_at_1000_std value: 18.805424261346374 - type: nauc_mrr_at_100_diff1 value: 43.55633267437543 - type: nauc_mrr_at_100_max value: 47.14214569591525 - type: nauc_mrr_at_100_std value: 18.828541893531277 - type: nauc_mrr_at_10_diff1 value: 43.07000882702881 - type: nauc_mrr_at_10_max value: 47.10398430807609 - type: nauc_mrr_at_10_std value: 18.672657418468155 - type: nauc_mrr_at_1_diff1 value: 50.71044015206451 - type: nauc_mrr_at_1_max value: 50.31094117388535 - type: nauc_mrr_at_1_std value: 16.308699760476404 - type: nauc_mrr_at_20_diff1 value: 43.34419341411509 - type: nauc_mrr_at_20_max value: 47.127839363881634 - type: nauc_mrr_at_20_std value: 18.93672383999524 - type: nauc_mrr_at_3_diff1 value: 44.09886232125989 - type: nauc_mrr_at_3_max value: 47.35761798607356 - type: nauc_mrr_at_3_std value: 18.66293179466984 - type: nauc_mrr_at_5_diff1 value: 43.455234122310486 - type: nauc_mrr_at_5_max value: 46.95579311628989 - type: nauc_mrr_at_5_std value: 18.637801785868913 - type: nauc_ndcg_at_1000_diff1 value: 42.09778197382488 - type: nauc_ndcg_at_1000_max value: 46.41254633930011 - type: nauc_ndcg_at_1000_std value: 19.727442899891408 - type: nauc_ndcg_at_100_diff1 value: 42.127587196947616 - type: nauc_ndcg_at_100_max value: 46.56257426488274 - type: nauc_ndcg_at_100_std value: 20.848893214507893 - type: nauc_ndcg_at_10_diff1 value: 39.520585737534184 - type: nauc_ndcg_at_10_max value: 45.58832499779741 - type: nauc_ndcg_at_10_std value: 19.230954524847657 - type: nauc_ndcg_at_1_diff1 value: 50.71044015206451 - type: nauc_ndcg_at_1_max value: 50.31094117388535 - type: nauc_ndcg_at_1_std value: 16.308699760476404 - type: nauc_ndcg_at_20_diff1 value: 40.57140695180754 - type: nauc_ndcg_at_20_max value: 45.78884507871275 - type: nauc_ndcg_at_20_std value: 20.87311919719877 - type: nauc_ndcg_at_3_diff1 value: 42.23214214323953 - type: nauc_ndcg_at_3_max value: 44.25227959403861 - type: nauc_ndcg_at_3_std value: 16.808716032720582 - type: nauc_ndcg_at_5_diff1 value: 40.32970262607426 - type: nauc_ndcg_at_5_max value: 44.170446333441234 - type: nauc_ndcg_at_5_std value: 17.670796157538952 - type: nauc_precision_at_1000_diff1 value: 4.4855757822300575 - type: nauc_precision_at_1000_max value: 40.96816841248859 - type: nauc_precision_at_1000_std value: 52.76450049154224 - type: nauc_precision_at_100_diff1 value: 13.467456291972423 - type: nauc_precision_at_100_max value: 46.07633674307899 - type: nauc_precision_at_100_std value: 58.38655747924394 - type: nauc_precision_at_10_diff1 value: 18.885447707274754 - type: nauc_precision_at_10_max value: 47.475287933169 - type: nauc_precision_at_10_std value: 40.78242836332111 - type: nauc_precision_at_1_diff1 value: 50.71044015206451 - type: nauc_precision_at_1_max value: 50.31094117388535 - type: nauc_precision_at_1_std value: 16.308699760476404 - type: nauc_precision_at_20_diff1 value: 15.953924273102402 - type: nauc_precision_at_20_max value: 45.47509365077202 - type: nauc_precision_at_20_std value: 51.47100789520174 - type: nauc_precision_at_3_diff1 value: 34.84717380734587 - type: nauc_precision_at_3_max value: 45.610933933265756 - type: nauc_precision_at_3_std value: 27.734101378690852 - type: nauc_precision_at_5_diff1 value: 26.59896898222078 - type: nauc_precision_at_5_max value: 46.140890589971264 - type: nauc_precision_at_5_std value: 33.56649457748371 - type: nauc_recall_at_1000_diff1 value: 86.92810457516407 - type: nauc_recall_at_1000_max value: 100.0 - type: nauc_recall_at_1000_std value: 100.0 - type: nauc_recall_at_100_diff1 value: 43.86702049240759 - type: nauc_recall_at_100_max value: 53.33308762101326 - type: nauc_recall_at_100_std value: 63.09523809523798 - type: nauc_recall_at_10_diff1 value: 25.88560487444265 - type: nauc_recall_at_10_max value: 41.6157709657381 - type: nauc_recall_at_10_std value: 24.04962076662668 - type: nauc_recall_at_1_diff1 value: 48.949793800671124 - type: nauc_recall_at_1_max value: 45.82239538118238 - type: nauc_recall_at_1_std value: 11.183927196674755 - type: nauc_recall_at_20_diff1 value: 27.507691414639822 - type: nauc_recall_at_20_max value: 41.70246318763185 - type: nauc_recall_at_20_std value: 37.33722257696256 - type: nauc_recall_at_3_diff1 value: 35.956192998402784 - type: nauc_recall_at_3_max value: 38.74690791289058 - type: nauc_recall_at_3_std value: 15.683526476441553 - type: nauc_recall_at_5_diff1 value: 31.03358342668625 - type: nauc_recall_at_5_max value: 37.820450291250786 - type: nauc_recall_at_5_std value: 18.52848795003198 - type: ndcg_at_1 value: 35.0 - type: ndcg_at_10 value: 52.842 - type: ndcg_at_100 value: 57.513999999999996 - type: ndcg_at_1000 value: 58.272999999999996 - type: ndcg_at_20 value: 55.454 - type: ndcg_at_3 value: 45.452 - type: ndcg_at_5 value: 49.169000000000004 - type: precision_at_1 value: 35.0 - type: precision_at_10 value: 8.366999999999999 - type: precision_at_100 value: 1.0630000000000002 - type: precision_at_1000 value: 0.11299999999999999 - type: precision_at_20 value: 4.75 - type: precision_at_3 value: 19.333 - type: precision_at_5 value: 14.066999999999998 - type: recall_at_1 value: 32.911 - type: recall_at_10 value: 73.033 - type: recall_at_100 value: 93.667 - type: recall_at_1000 value: 99.667 - type: recall_at_20 value: 83.0 - type: recall_at_3 value: 52.878 - type: recall_at_5 value: 62.06700000000001 - task: type: PairClassification dataset: name: MTEB SprintDuplicateQuestions (default) type: mteb/sprintduplicatequestions-pairclassification config: default split: test revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46 metrics: - type: cosine_accuracy value: 99.87425742574257 - type: cosine_accuracy_threshold value: 85.4932188987732 - type: cosine_ap value: 97.03588351132844 - type: cosine_f1 value: 93.60201511335012 - type: cosine_f1_threshold value: 85.4932188987732 - type: cosine_precision value: 94.31472081218274 - type: cosine_recall value: 92.9 - type: dot_accuracy value: 99.87425742574257 - type: dot_accuracy_threshold value: 85.4932188987732 - type: dot_ap value: 97.03588351132846 - type: dot_f1 value: 93.60201511335012 - type: dot_f1_threshold value: 85.4932188987732 - type: dot_precision value: 94.31472081218274 - type: dot_recall value: 92.9 - type: euclidean_accuracy value: 99.87425742574257 - type: euclidean_accuracy_threshold value: 53.864240646362305 - type: euclidean_ap value: 97.03588351132844 - type: euclidean_f1 value: 93.60201511335012 - type: euclidean_f1_threshold value: 53.864240646362305 - type: euclidean_precision value: 94.31472081218274 - type: euclidean_recall value: 92.9 - type: main_score value: 97.12020380643673 - type: manhattan_accuracy value: 99.87821782178217 - type: manhattan_accuracy_threshold value: 2557.1868896484375 - type: manhattan_ap value: 97.12020380643673 - type: manhattan_f1 value: 93.83458646616543 - type: manhattan_f1_threshold value: 2559.8316192626953 - type: manhattan_precision value: 94.07035175879398 - type: manhattan_recall value: 93.60000000000001 - type: max_accuracy value: 99.87821782178217 - type: max_ap value: 97.12020380643673 - type: max_f1 value: 93.83458646616543 - type: max_precision value: 94.31472081218274 - type: max_recall value: 93.60000000000001 - type: similarity_accuracy value: 99.87425742574257 - type: similarity_accuracy_threshold value: 85.4932188987732 - type: similarity_ap value: 97.03588351132844 - type: similarity_f1 value: 93.60201511335012 - type: similarity_f1_threshold value: 85.4932188987732 - type: similarity_precision value: 94.31472081218274 - type: similarity_recall value: 92.9 - task: type: Clustering dataset: name: MTEB StackExchangeClustering (default) type: mteb/stackexchange-clustering config: default split: test revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259 metrics: - type: main_score value: 76.98818225336838 - type: v_measure value: 76.98818225336838 - type: v_measure_std value: 3.154967965946174 - task: type: Clustering dataset: name: MTEB StackExchangeClusteringP2P (default) type: mteb/stackexchange-clustering-p2p config: default split: test revision: 815ca46b2622cec33ccafc3735d572c266efdb44 metrics: - type: main_score value: 45.163651140607605 - type: v_measure value: 45.163651140607605 - type: v_measure_std value: 1.4322970276083837 - task: type: Reranking dataset: name: MTEB StackOverflowDupQuestions (default) type: mteb/stackoverflowdupquestions-reranking config: default split: test revision: e185fbe320c72810689fc5848eb6114e1ef5ec69 metrics: - type: main_score value: 56.391883714372696 - type: map value: 56.391883714372696 - type: mrr value: 57.349492827434 - type: nAUC_map_diff1 value: 39.157250127064955 - type: nAUC_map_max value: 18.467392575309553 - type: nAUC_map_std value: 6.562904741623687 - type: nAUC_mrr_diff1 value: 39.2616391317946 - type: nAUC_mrr_max value: 20.17824080849778 - type: nAUC_mrr_std value: 7.3151994802766005 - task: type: Summarization dataset: name: MTEB SummEval (default) type: mteb/summeval config: default split: test revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c metrics: - type: cosine_pearson value: 31.115370364087013 - type: cosine_spearman value: 30.168250595399797 - type: dot_pearson value: 31.11537534713581 - type: dot_spearman value: 30.168250595399797 - type: main_score value: 30.168250595399797 - type: pearson value: 31.115370364087013 - type: spearman value: 30.168250595399797 - task: type: Retrieval dataset: name: MTEB TRECCOVID (default) type: mteb/trec-covid config: default split: test revision: bb9466bac8153a0349341eb1b22e06409e78ef4e metrics: - type: main_score value: 58.492 - type: map_at_1 value: 0.20600000000000002 - type: map_at_10 value: 1.355 - type: map_at_100 value: 7.682 - type: map_at_1000 value: 19.422 - type: map_at_20 value: 2.307 - type: map_at_3 value: 0.504 - type: map_at_5 value: 0.756 - type: mrr_at_1 value: 76.0 - type: mrr_at_10 value: 83.07460317460317 - type: mrr_at_100 value: 83.34653299916457 - type: mrr_at_1000 value: 83.34653299916457 - type: mrr_at_20 value: 83.34653299916457 - type: mrr_at_3 value: 81.66666666666666 - type: mrr_at_5 value: 82.56666666666666 - type: nauc_map_at_1000_diff1 value: -0.9517122101342602 - type: nauc_map_at_1000_max value: 35.489825727736665 - type: nauc_map_at_1000_std value: 72.31927320292716 - type: nauc_map_at_100_diff1 value: -2.6696855309157197 - type: nauc_map_at_100_max value: 16.881012905948 - type: nauc_map_at_100_std value: 60.636797544764796 - type: nauc_map_at_10_diff1 value: 3.3220618387062166 - type: nauc_map_at_10_max value: 7.9728051776655136 - type: nauc_map_at_10_std value: 37.001872811447676 - type: nauc_map_at_1_diff1 value: 19.385947791364455 - type: nauc_map_at_1_max value: -2.017784609408856 - type: nauc_map_at_1_std value: 15.846915472515105 - type: nauc_map_at_20_diff1 value: 1.0613460412567055 - type: nauc_map_at_20_max value: 7.639419874542262 - type: nauc_map_at_20_std value: 42.004875229740826 - type: nauc_map_at_3_diff1 value: 7.0015165243253366 - type: nauc_map_at_3_max value: 7.084211457521959 - type: nauc_map_at_3_std value: 24.788352390570584 - type: nauc_map_at_5_diff1 value: 6.657899114095232 - type: nauc_map_at_5_max value: 4.976947597730104 - type: nauc_map_at_5_std value: 29.481454683941184 - type: nauc_mrr_at_1000_diff1 value: 14.561577730498792 - type: nauc_mrr_at_1000_max value: 57.72810732532122 - type: nauc_mrr_at_1000_std value: 66.88388647529588 - type: nauc_mrr_at_100_diff1 value: 14.561577730498792 - type: nauc_mrr_at_100_max value: 57.72810732532122 - type: nauc_mrr_at_100_std value: 66.88388647529588 - type: nauc_mrr_at_10_diff1 value: 14.57469254485188 - type: nauc_mrr_at_10_max value: 58.079825098428714 - type: nauc_mrr_at_10_std value: 67.32128458796227 - type: nauc_mrr_at_1_diff1 value: 25.34827377347056 - type: nauc_mrr_at_1_max value: 50.58838798996285 - type: nauc_mrr_at_1_std value: 59.36661763433414 - type: nauc_mrr_at_20_diff1 value: 14.561577730498792 - type: nauc_mrr_at_20_max value: 57.72810732532122 - type: nauc_mrr_at_20_std value: 66.88388647529588 - type: nauc_mrr_at_3_diff1 value: 9.063532868160214 - type: nauc_mrr_at_3_max value: 58.71832537642312 - type: nauc_mrr_at_3_std value: 69.07730444362834 - type: nauc_mrr_at_5_diff1 value: 13.555968426927894 - type: nauc_mrr_at_5_max value: 59.22085120600723 - type: nauc_mrr_at_5_std value: 67.47575721875769 - type: nauc_ndcg_at_1000_diff1 value: -1.8751322983265282 - type: nauc_ndcg_at_1000_max value: 38.78712823179003 - type: nauc_ndcg_at_1000_std value: 70.43132053994896 - type: nauc_ndcg_at_100_diff1 value: -10.220936212671377 - type: nauc_ndcg_at_100_max value: 47.70220514113511 - type: nauc_ndcg_at_100_std value: 75.65229647100806 - type: nauc_ndcg_at_10_diff1 value: 2.0956279601914227 - type: nauc_ndcg_at_10_max value: 48.868693823231304 - type: nauc_ndcg_at_10_std value: 70.16734895474447 - type: nauc_ndcg_at_1_diff1 value: 27.89880129091742 - type: nauc_ndcg_at_1_max value: 44.14668818195789 - type: nauc_ndcg_at_1_std value: 60.28699861687413 - type: nauc_ndcg_at_20_diff1 value: -3.5946895305356623 - type: nauc_ndcg_at_20_max value: 46.68859141418255 - type: nauc_ndcg_at_20_std value: 70.27067652865686 - type: nauc_ndcg_at_3_diff1 value: 7.400409149522286 - type: nauc_ndcg_at_3_max value: 45.61078758588923 - type: nauc_ndcg_at_3_std value: 62.06453130401961 - type: nauc_ndcg_at_5_diff1 value: 5.830725665736509 - type: nauc_ndcg_at_5_max value: 46.62678021725239 - type: nauc_ndcg_at_5_std value: 64.28848314363539 - type: nauc_precision_at_1000_diff1 value: -9.666313428844905 - type: nauc_precision_at_1000_max value: 47.57616298626001 - type: nauc_precision_at_1000_std value: 49.81803250713608 - type: nauc_precision_at_100_diff1 value: -10.753663329125686 - type: nauc_precision_at_100_max value: 45.231033820687834 - type: nauc_precision_at_100_std value: 74.22025319558313 - type: nauc_precision_at_10_diff1 value: -0.9044324563451003 - type: nauc_precision_at_10_max value: 46.282938258557955 - type: nauc_precision_at_10_std value: 67.20654075066248 - type: nauc_precision_at_1_diff1 value: 25.34827377347056 - type: nauc_precision_at_1_max value: 50.58838798996285 - type: nauc_precision_at_1_std value: 59.36661763433414 - type: nauc_precision_at_20_diff1 value: -5.192190687520166 - type: nauc_precision_at_20_max value: 39.61181596936397 - type: nauc_precision_at_20_std value: 65.90673204251821 - type: nauc_precision_at_3_diff1 value: -1.1581585542804733 - type: nauc_precision_at_3_max value: 48.095238095238116 - type: nauc_precision_at_3_std value: 57.79976256430543 - type: nauc_precision_at_5_diff1 value: 3.355915932928888 - type: nauc_precision_at_5_max value: 43.99987410397438 - type: nauc_precision_at_5_std value: 62.106083138587906 - type: nauc_recall_at_1000_diff1 value: 3.655993902820825 - type: nauc_recall_at_1000_max value: 28.761919544640335 - type: nauc_recall_at_1000_std value: 61.94123910402753 - type: nauc_recall_at_100_diff1 value: 2.5155941410242977 - type: nauc_recall_at_100_max value: 9.499702402437284 - type: nauc_recall_at_100_std value: 52.57449917231589 - type: nauc_recall_at_10_diff1 value: 5.939411921276368 - type: nauc_recall_at_10_max value: 4.994244760738587 - type: nauc_recall_at_10_std value: 33.64383950012248 - type: nauc_recall_at_1_diff1 value: 19.385947791364455 - type: nauc_recall_at_1_max value: -2.017784609408856 - type: nauc_recall_at_1_std value: 15.846915472515105 - type: nauc_recall_at_20_diff1 value: 3.339213533105717 - type: nauc_recall_at_20_max value: 1.4182715611821584 - type: nauc_recall_at_20_std value: 36.13152761959804 - type: nauc_recall_at_3_diff1 value: 2.9154975009752775 - type: nauc_recall_at_3_max value: 5.418186566728512 - type: nauc_recall_at_3_std value: 24.420940449950507 - type: nauc_recall_at_5_diff1 value: 7.4799616256209305 - type: nauc_recall_at_5_max value: 2.1601588551873823 - type: nauc_recall_at_5_std value: 28.09415304774757 - type: ndcg_at_1 value: 72.0 - type: ndcg_at_10 value: 58.492 - type: ndcg_at_100 value: 45.437 - type: ndcg_at_1000 value: 44.108999999999995 - type: ndcg_at_20 value: 54.969 - type: ndcg_at_3 value: 64.93900000000001 - type: ndcg_at_5 value: 60.736999999999995 - type: precision_at_1 value: 76.0 - type: precision_at_10 value: 61.199999999999996 - type: precision_at_100 value: 46.839999999999996 - type: precision_at_1000 value: 19.666 - type: precision_at_20 value: 56.8 - type: precision_at_3 value: 68.0 - type: precision_at_5 value: 62.8 - type: recall_at_1 value: 0.20600000000000002 - type: recall_at_10 value: 1.5939999999999999 - type: recall_at_100 value: 11.498 - type: recall_at_1000 value: 42.729 - type: recall_at_20 value: 2.922 - type: recall_at_3 value: 0.5309999999999999 - type: recall_at_5 value: 0.8370000000000001 - task: type: Classification dataset: name: MTEB ToxicConversationsClassification (default) type: mteb/toxic_conversations_50k config: default split: test revision: edfaf9da55d3dd50d43143d90c1ac476895ae6de metrics: - type: accuracy value: 85.9521484375 - type: ap value: 30.374730390938566 - type: ap_weighted value: 30.374730390938566 - type: f1 value: 70.3917271343218 - type: f1_weighted value: 88.45609971763992 - type: main_score value: 85.9521484375 - task: type: Classification dataset: name: MTEB TweetSentimentExtractionClassification (default) type: mteb/tweet_sentiment_extraction config: default split: test revision: d604517c81ca91fe16a244d1248fc021f9ecee7a metrics: - type: accuracy value: 80.12733446519525 - type: f1 value: 80.418094849412 - type: f1_weighted value: 80.10847441279616 - type: main_score value: 80.12733446519525 - task: type: Clustering dataset: name: MTEB TwentyNewsgroupsClustering (default) type: mteb/twentynewsgroups-clustering config: default split: test revision: 6125ec4e24fa026cec8a478383ee943acfbd5449 metrics: - type: main_score value: 64.6036121602603 - type: v_measure value: 64.6036121602603 - type: v_measure_std value: 1.2991377356017484 - task: type: PairClassification dataset: name: MTEB TwitterSemEval2015 (default) type: mteb/twittersemeval2015-pairclassification config: default split: test revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1 metrics: - type: cosine_accuracy value: 87.86433808189784 - type: cosine_accuracy_threshold value: 85.5525255203247 - type: cosine_ap value: 78.93155350890012 - type: cosine_f1 value: 71.80031864046734 - type: cosine_f1_threshold value: 83.99585485458374 - type: cosine_precision value: 72.26082308925709 - type: cosine_recall value: 71.34564643799473 - type: dot_accuracy value: 87.86433808189784 - type: dot_accuracy_threshold value: 85.55253744125366 - type: dot_ap value: 78.93157147282707 - type: dot_f1 value: 71.80031864046734 - type: dot_f1_threshold value: 83.99585485458374 - type: dot_precision value: 72.26082308925709 - type: dot_recall value: 71.34564643799473 - type: euclidean_accuracy value: 87.86433808189784 - type: euclidean_accuracy_threshold value: 53.75403165817261 - type: euclidean_ap value: 78.93157128337329 - type: euclidean_f1 value: 71.80031864046734 - type: euclidean_f1_threshold value: 56.575870513916016 - type: euclidean_precision value: 72.26082308925709 - type: euclidean_recall value: 71.34564643799473 - type: main_score value: 79.12654131533807 - type: manhattan_accuracy value: 87.98950944745782 - type: manhattan_accuracy_threshold value: 2512.5680923461914 - type: manhattan_ap value: 79.12654131533807 - type: manhattan_f1 value: 71.90745366110163 - type: manhattan_f1_threshold value: 2624.722671508789 - type: manhattan_precision value: 71.65313073094053 - type: manhattan_recall value: 72.16358839050132 - type: max_accuracy value: 87.98950944745782 - type: max_ap value: 79.12654131533807 - type: max_f1 value: 71.90745366110163 - type: max_precision value: 72.26082308925709 - type: max_recall value: 72.16358839050132 - type: similarity_accuracy value: 87.86433808189784 - type: similarity_accuracy_threshold value: 85.5525255203247 - type: similarity_ap value: 78.93155350890012 - type: similarity_f1 value: 71.80031864046734 - type: similarity_f1_threshold value: 83.99585485458374 - type: similarity_precision value: 72.26082308925709 - type: similarity_recall value: 71.34564643799473 - task: type: PairClassification dataset: name: MTEB TwitterURLCorpus (default) type: mteb/twitterurlcorpus-pairclassification config: default split: test revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf metrics: - type: cosine_accuracy value: 89.03248340901153 - type: cosine_accuracy_threshold value: 84.39068794250488 - type: cosine_ap value: 85.87150718008797 - type: cosine_f1 value: 78.39147286821706 - type: cosine_f1_threshold value: 82.88650512695312 - type: cosine_precision value: 75.96792834440913 - type: cosine_recall value: 80.97474591931014 - type: dot_accuracy value: 89.03248340901153 - type: dot_accuracy_threshold value: 84.39069986343384 - type: dot_ap value: 85.87150946221163 - type: dot_f1 value: 78.39147286821706 - type: dot_f1_threshold value: 82.88650512695312 - type: dot_precision value: 75.96792834440913 - type: dot_recall value: 80.97474591931014 - type: euclidean_accuracy value: 89.03248340901153 - type: euclidean_accuracy_threshold value: 55.873626470565796 - type: euclidean_ap value: 85.87151445202907 - type: euclidean_f1 value: 78.39147286821706 - type: euclidean_f1_threshold value: 58.5038423538208 - type: euclidean_precision value: 75.96792834440913 - type: euclidean_recall value: 80.97474591931014 - type: main_score value: 85.95871260636034 - type: manhattan_accuracy value: 89.09069740365584 - type: manhattan_accuracy_threshold value: 2603.150749206543 - type: manhattan_ap value: 85.95871260636034 - type: manhattan_f1 value: 78.53649430651484 - type: manhattan_f1_threshold value: 2714.5809173583984 - type: manhattan_precision value: 76.23396390519677 - type: manhattan_recall value: 80.9824453341546 - type: max_accuracy value: 89.09069740365584 - type: max_ap value: 85.95871260636034 - type: max_f1 value: 78.53649430651484 - type: max_precision value: 76.23396390519677 - type: max_recall value: 80.9824453341546 - type: similarity_accuracy value: 89.03248340901153 - type: similarity_accuracy_threshold value: 84.39068794250488 - type: similarity_ap value: 85.87150718008797 - type: similarity_f1 value: 78.39147286821706 - type: similarity_f1_threshold value: 82.88650512695312 - type: similarity_precision value: 75.96792834440913 - type: similarity_recall value: 80.97474591931014 ---
[ "BIOSSES", "SCIFACT" ]
deadman44/SDXL_Photoreal_Merged_Models
deadman44
text-to-image
[ "text-to-image", "stable-diffusion", "safetensors", "stable-diffusion-xl", "en", "license:other", "region:us" ]
"2024-01-18T10:43:23Z"
2025-03-12T05:05:33+00:00
0
62
--- language: - en license: other license_name: faipl-1.0-sd license_link: https://freedevproject.org/faipl-1.0-sd/ tags: - text-to-image - stable-diffusion - safetensors - stable-diffusion-xl --- <style> .title{ font-size: 2.5em; letter-spacing: 0.01em; padding: 0.5em 0; } .thumbwidth{ max-width: 180px; } .font_red{ color:red; } .font_blue{ color:blue; } .font_grey{ color: #aaaaaa; } </style> # models - [Zipatrious_XL_v1.0](#zipatrious1) (<span class="font_red">Illustrious Base</span>):2025-03-12<br /> - [Zipanoob_XL_Epred_v1.1](#zipanoob1) (<span class="font_red">Noob AI Base</span>):2025-01-31<br /> - [Zipanoob_XL_Vpred_v1.1](#zipanoob1) (<span class="font_red">Noob AI Base</span>):2025-01-14<br /> - [Zipang XL test03.1](#test031) (<span class="font_red">Animagine Base</span>): 3.1b:2024-07-22<br /> - [Ponypang_XL_giveup](#potest2) (<span class="font_red">Pony Base</span>): fix5:2024-07-03<br /> --- <a id="zipatrious1"></a> <h1 class="title"> <span>Zipatrious XL v1.0</span> </h1> -20000+ images Finetune trained<br/> -<span class="font_red">Experimental Version</span><br/> -<span class="font_blue">More realistic</span><br/> -<span class="font_red">Expression is not wide</span><br/> -<span class="font_red">Many bankruptcies</span><br/> -<span class="font_red">Animated characters cannot be output.</span><br/> <br/> -<a href="https://huggingface.co/tianweiy/DMD2/blob/main/dmd2_sdxl_4step_lora.safetensors">dmd2_sdxl_4step_lora</a> included<br/> -<span class="font_red">ADetailer or HiresFix Recommendation.</span><br/><br/> <br/> <br/> # Recommendation <span class="font_blue">Euler a Uniform 10 steps CFG Scale:1.3</span> <br/> <br/> [Download: v1.0](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/Zipatrious_XL_v1.0.safetensors?download=true)<br/> <br/> # - base model -<a href="https://huggingface.co/OnomaAIResearch/Illustrious-xl-early-release-v0"> Illustrious-xl-early-release-v0</a><br/> -great XL model!<br /> <br /> ## - trigger ```bash japanese european 3yo-30yo myob, myjd, myjk, myjc, myjsh, myjsm, myjsl, myjy (one of these) ``` <br /> ## - quality tags ```bash masterpiece, best quality, realistic, photorealistic, ``` <br /> ## - negative tags ```bash low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits ``` <br /> ## - prompt ```bash danbooru tag + natural english ``` <br /> ## - Sampling method ```bash Euler a :10 steps ``` <br /> ## - Schedule type ```bash Uniform ``` <br /> ## - cfg scale ```bash 1-1.5 ``` <br /> ## - Resolution ```bash 768 x1024, 1024 x 768 1024 x 1024 896 x 1152, 1152 x 896 512 x 768, 768 x 512 832 x 1216 ``` <br /> ## - Recommended ```bash -Hires.fix: Hires steps:3, Denoising strength:0.1-0.2, Hires CFG Scale:1.3 or ADetailer: yoroface, Inpaint denoising strength:0.1-0.3 + i2i(same settings t2i, Denoising strength:0.1-0.3) + script: Ultra SD upscale (*external script) ``` <br /> --- # sample prompt [<img src=https://t12.pixhost.to/thumbs/866/575299674_20250312131222_zipatrious_xl_v1-0_1747029599.jpg />](https://img12.pixhost.to/images/866/575299674_20250312131222_zipatrious_xl_v1-0_1747029599.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, 25yo, myob, japanese, 1girl, solo, black hair, long hair, black eyes, closed mouth, looking at viewer, smile, sitting, arm support, on floor, sweater dress, ribbed sweater, sleeves past wrists, denim long pants, knee up, leaning back, hand on leg, indoors, sunlight, day, window, curtains, backlighting, windowsill, balcony, feet out of frame Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 10, Sampler: Euler a, Schedule type: Uniform, CFG scale: 1.3, Seed: 1747029599, Size: 832x1216, Model hash: 1699514967, Model: Zipatrious_XL_v1.0, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t12.pixhost.to/thumbs/866/575299681_20250312132311_zipatrious_xl_v1-0_1091038412.jpg />](https://img12.pixhost.to/images/866/575299681_20250312132311_zipatrious_xl_v1-0_1091038412.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, 19yo, myjd, japanese, 1girl, squatting, brown hair, smile, parted lips, brown eyes, shoulder bag, long hair, ponytail, looking at viewer, white shirt, dress shirt, black skirt, shadow, night, back alley, photo background, close-up Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 10, Sampler: Euler a, Schedule type: Uniform, CFG scale: 1.3, Seed: 1091038412, Size: 832x1216, Model hash: 1699514967, Model: Zipatrious_XL_v1.0, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t12.pixhost.to/thumbs/866/575299687_20250312133407_zipatrious_xl_v1-0_1089820708.jpg />](https://img12.pixhost.to/images/866/575299687_20250312133407_zipatrious_xl_v1-0_1089820708.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, 12yo, myjsh, japanese, 1girl, solo, black casual hair, lying on couch, green couch, medium skirt, wooden room, white casual dress, long sleeve, detailed, grin, looking at viewer Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 10, Sampler: Euler a, Schedule type: Uniform, CFG scale: 1.3, Seed: 1089820708, Size: 832x1216, Model hash: 1699514967, Model: Zipatrious_XL_v1.0, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t12.pixhost.to/thumbs/866/575299694_20250312134416_zipatrious_xl_v1-0_3234313253.jpg />](https://img12.pixhost.to/images/866/575299694_20250312134416_zipatrious_xl_v1-0_3234313253.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, 28yo, myob, european, 1girl, mature female, solo, long wavy hair, blonde hair, blue eyes, smile, salute, air force uniform, pilot, sceanery, blurry, cowboy shot Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 10, Sampler: Euler a, Schedule type: Uniform, CFG scale: 1.3, Seed: 3234313253, Size: 832x1216, Model hash: 1699514967, Model: Zipatrious_XL_v1.0, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> <br /> --- <a id="zipanoob1"></a> <h1 class="title"> <span>Zipanoob XL Vpred v1.1</span> </h1> -20000+ images Finetune trained<br/> <br/> -(Epred) Base model:NoobAI XL Epsilon-pred 1.1-Version<br/> -(Vpred) Base model:NoobAI XL V-Pred 1.0<br/> -<a href="https://huggingface.co/tianweiy/DMD2/blob/main/dmd2_sdxl_4step_lora.safetensors">dmd2_sdxl_4step_lora</a> included<br/> -<span class="font_red">ADetailer or HiresFix Recommendation.</span><br/><br/> ***<span class="font_red">(Vpred version)</span><br/> -(forge)<span class="font_blue">turn on</span> Zero Terminal SNR<br/> ***<span class="font_blue">Turn off</span> when using flux or other epred models<br/> <br/> <br/> [Download: Epred_v1.1](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/Zipanoob_XL_Epred_v1.1.safetensors?download=true)<br/> [Download: Vpred_v1.1](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/Zipanoob_XL_Vpred_v1.1.safetensors?download=true)<br/> <br/> # - base model -<a href="https://civitai.com/models/833294?modelVersionId=1116447">NoobAI XL Epsilon-pred 1.1-Version</a><br/> -<a href="https://civitai.com/models/833294?modelVersionId=1190596">NoobAI XL V-Pred 1.0</a><br/> -great XL model!<br /> <br /> ## - trigger ```bash japanese european 3yo-30yo myob, myjd, myjk, myjc, myjsh, myjsm, myjsl, myjy (one of these) ``` <br /> ## - quality tags ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, ``` <br /> ## - negative tags ```bash low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits ``` <br /> ## - prompt ```bash danbooru tag + natural english ``` <br /> ## - Sampling method ```bash Euler a :6 steps ``` <br /> ## - Schedule type ```bash SGM Uniform, KL Optimal, Normal, Simple ``` <br /> ## - cfg scale ```bash 1-1.5 ``` <br /> ## - Resolution ```bash 768 x1024, 1024 x 768 1024 x 1024 896 x 1152, 1152 x 896 512 x 768, 768 x 512 ``` <br /> ## - Recommended ```bash -Hires.fix: Hires steps:3, Denoising strength:0.1-0.2, Hires CFG Scale:1.3 or ADetailer: yoroface, Inpaint denoising strength:0.1-0.3 + i2i(same settings t2i, Denoising strength:0.1-0.3) + script: Ultra SD upscale (*external script) ``` <br /> --- # sample prompt (Image is Vpred version) [<img src=https://t101.pixhost.to/thumbs/558/555277222_20250114134415_zipanoob_xl_vpred_v1-1_429389538.jpg />](https://img101.pixhost.to/images/558/555277222_20250114134415_zipanoob_xl_vpred_v1-1_429389538.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 19yo, myjd, japanese, 1girl, solo, black hair, long hair, brown eyes, makeup, lipstick, smile, closed mouth, white dress, lace-trimmed dress, lace trim, laceshort sleeves, sleeveless, looking at viewer, squatting, knees up, depth of field, lens flare Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 429389538, Size: 896x1152, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Hires Module 1: Use same choices, Hires CFG Scale: 1.3, Hires upscale: 2, Hires steps: 3, Hires upscaler: 4x-UltraSharp, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/558/555278714_20250114140909_zipanoob_xl_vpred_v1-1_1091415239.jpg />](https://img101.pixhost.to/images/558/555278714_20250114140909_zipanoob_xl_vpred_v1-1_1091415239.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, myjk, japanese, 2girls, multiple girls, black hair, long hair, short hair, ponytail, bangs, brown eyes, school uniform, white shirt, short sleeves, bowtie, smile, happy, looking at viewer, selfie, v, train interior, window Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 1091415239, Size: 1792x2304, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Ultimate SD upscale upscaler: 4x-UltraSharp, Ultimate SD upscale tile_width: 896, Ultimate SD upscale tile_height: 1152, Ultimate SD upscale mask_blur: 8, Ultimate SD upscale padding: 32, Mask blur: 8, Inpaint area: Only masked, Masked area padding: 32, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/558/555278715_20250114142203_zipanoob_xl_vpred_v1-1_3192029900.jpg />](https://img101.pixhost.to/images/558/555278715_20250114142203_zipanoob_xl_vpred_v1-1_3192029900.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 15yo, myjc, japanese, 1girl, solo, very long hair, black hair, hair bow, black eyes, blue dress, detached sleeves, wide sleeves, bare shoulders, blue skirt, fur-trimmed skirt, white over-kneehighs, white socks, boots, grin, holding microphone, standing, looking at viewer, hand up, live stage, spot lighting, shadow, cowboy shot, winf Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 3192029900, Size: 896x1152, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.2, Hires Module 1: Use same choices, Hires CFG Scale: 1.3, Hires upscale: 2, Hires steps: 3, Hires upscaler: 4x-UltraSharp, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/559/555296170_20250114163258_zipanoob_xl_vpred_v1-1_2021607063.jpg />](https://img101.pixhost.to/images/559/555296170_20250114163258_zipanoob_xl_vpred_v1-1_2021607063.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 24yo, myob, japanese, 1girl, solo focus, crowd, Akihabara, maid, buissines suit, day lighting, wind, dutch angle Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 2021607063, Size: 1792x2304, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Ultimate SD upscale upscaler: 4x-UltraSharp, Ultimate SD upscale tile_width: 896, Ultimate SD upscale tile_height: 1152, Ultimate SD upscale mask_blur: 8, Ultimate SD upscale padding: 32, Mask blur: 8, Inpaint area: Only masked, Masked area padding: 32, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/559/555296176_20250114165441_zipanoob_xl_vpred_v1-1_197566853.jpg />](https://img101.pixhost.to/images/560/555302145_20250114165441_zipanoob_xl_vpred_v1-1_197566853.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 23yo, myob, european, woman, solo, blonde hair, looking at viewer, grin, blue eyes, wavy hair, space ship, sitting, knee up, crossed legs, windw, earth \(planet\), head tilt, cowboy shot, feet out of frame, milf, astronaut Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 197566853, Size: 896x1152, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.3, Hires Module 1: Use same choices, Hires CFG Scale: 1.3, Hires upscale: 2, Hires steps: 3, Hires upscaler: 4x-UltraSharp, Emphasis: No norm, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/559/555296169_20250114155644_zipanoob_xl_vpred_v1-1_4151704591.jpg />](https://img101.pixhost.to/images/559/555296169_20250114155644_zipanoob_xl_vpred_v1-1_4151704591.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 27yo, myob, european, sunset, woman, ocean, outdoors, standing, backlighting, blonde hair, blue eyes, long hair, silhouette, shadow, horizon, white dress, wind, smile, upper body, adjusting Hair, looking to the side, fence Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 4151704591, Size: 896x1152, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.2, Hires Module 1: Use same choices, Hires CFG Scale: 1.3, Hires upscale: 2, Hires steps: 3, Hires upscaler: R-ESRGAN 4x+, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- ## Anime Characters (danbooru tag) [<img src=https://t101.pixhost.to/thumbs/558/555278718_20250114143709_zipanoob_xl_vpred_v1-1_3217608951.jpg />](https://img101.pixhost.to/images/558/555278718_20250114143709_zipanoob_xl_vpred_v1-1_3217608951.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, 10yo, myjsm, japanese, 1girl, solo, zundamon, white shirt, green overall, smile, looking at viewer, cowboy shot, outdoors, hands on own waist Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 3217608951, Size: 1792x2304, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Ultimate SD upscale upscaler: 4x-UltraSharp, Ultimate SD upscale tile_width: 896, Ultimate SD upscale tile_height: 1152, Ultimate SD upscale mask_blur: 8, Ultimate SD upscale padding: 32, Mask blur: 8, Inpaint area: Only masked, Masked area padding: 32, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> --- [<img src=https://t101.pixhost.to/thumbs/558/555278721_20250114144655_zipanoob_xl_vpred_v1-1_2061259503.jpg />](https://img101.pixhost.to/images/558/555278721_20250114144655_zipanoob_xl_vpred_v1-1_2061259503.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, myjc, european, frieren, fern \(sousou no frieren\), cosplay, 2girls, multiple girls, posing, at night, flash lighting, grin, looking at viewer, side-by-side Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 2061259503, Size: 1792x2304, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Ultimate SD upscale upscaler: 4x-UltraSharp, Ultimate SD upscale tile_width: 896, Ultimate SD upscale tile_height: 1152, Ultimate SD upscale mask_blur: 8, Ultimate SD upscale padding: 32, Mask blur: 8, Inpaint area: Only masked, Masked area padding: 32, Emphasis: No norm, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> [<img src=https://t101.pixhost.to/thumbs/558/555278723_20250114150144_zipanoob_xl_vpred_v1-1_1934868269.jpg />](https://img101.pixhost.to/images/558/555278723_20250114150144_zipanoob_xl_vpred_v1-1_1934868269.jpg) ```bash masterpiece, best quality, realistic, photorealistic, mole, body freckles, body mole, yoru \(spy x family\), anya \(spy x family\), 2girls, multiple girls, smile, sitting on chair, black uniform, natural lighting, looking at viewer, kitchen, side-by-side Negative prompt: low quality, worst quality, lowres, bad, bad anatomy, bad hands, multiple hands, mutation hands, fewer, extra, missing, displeasing, extra digits Steps: 6, Sampler: Euler a, Schedule type: SGM Uniform, CFG scale: 1.3, Seed: 1934868269, Size: 896x1152, Model hash: 2898567825, Model: Zipanoob_XL_Vpred_v1.1, Denoising strength: 0.1, Hires Module 1: Use same choices, Hires CFG Scale: 1.3, Hires upscale: 2, Hires steps: 3, Hires upscaler: R-ESRGAN 4x+, Emphasis: No norm, Noise Schedule: Zero Terminal SNR, Version: f2.0.1v1.10.1-previous-634-g37301b22, Module 1: sdxl.vae ``` <br /> <br /> --- ## -Train Settings - [sd-scripts (SD3 branch)](https://github.com/kohya-ss/sd-scripts/tree/sd3)<br> ```bash base model: NoobAI-XL-Vpred-v1.0.safetensors caption: JoyCaption Alpha Two tags: WD EVA02-Large Tagger v3 --network_module "sdxl_train.py" ^ --caption_dropout_rate="0" ^ --vae_batch_size="1" ^ --gradient_checkpointing ^ --persistent_data_loader_workers ^ --cache_latents ^ --cache_latents_to_disk ^ --max_data_loader_n_workers=2 ^ --enable_bucket ^ --bucket_no_upscale ^ --save_model_as "safetensors" ^ --mixed_precision "bf16" ^ --learning_rate=5e-6 ^ --train_text_encoder ^ --learning_rate_te1=5e-7 ^ --learning_rate_te2=5e-7 ^ --resolution=1024,1024 ^ --train_batch_size 2 ^ --optimizer_type "adafactor" ^ --optimizer_args "relative_step=False" "scale_parameter=False" "warmup_init=False" ^ --lr_scheduler "constant_with_warmup" ^ --save_precision "bf16" ^ --caption_extension ".txt" ^ --no_half_vae ^ --full_bf16 ^ --max_grad_norm=0 ^ --min_snr_gamma=5 ^ --max_token_length=225 ^ --fused_backward_pass ^ //Epred version --noise_offset=0.0375 ^ --adaptive_noise_scale=0.00375 ^ // //Vpred version --v_parameterization ^ --zero_terminal_snr ^ // --save_state ^ --xformers ``` <br /> --- <br /> <a id="test031"></a> <h1 class="title"> <span>Zipang XL test3.1</span> </h1> -4000+ twitter images trained & 10000+ images merged model<br/> <br/> -Animagine XL 3.1 Base<br/> -Good Lighting<br/> -photoreal like tag: shadow, flash lighting, backlighting, silhouette, sunset, night, day, bokeh, etc.<br/> <br/> [Download:test3.1](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/zipang_XL_test3.1.fp16.safetensors?download=true) (Recommended)<br/>[Download:test3.1b](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/zipang_XL_test3.1b.fp16.safetensors?download=true) (Newer isn't always better) <br/> <br/> These images are test3.1. <table> <tr> <td> <a href="https://img95.pixhost.to/images/1089/472649581_20240528114225_zipang_xl_test3-1-fp16_1076069870.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1089/472649581_20240528114225_zipang_xl_test3-1-fp16_1076069870.jpg" alt="sample1" class="thumbwidth" > </div> </a> <a href="https://img95.pixhost.to/images/1089/472649589_20240528120137_zipang_xl_test3-1-fp16_1729913506.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1089/472649589_20240528120137_zipang_xl_test3-1-fp16_1729913506.jpg" alt="sample2" class="thumbwidth" > </div> </a> </td> <td> <a href="https://img95.pixhost.to/images/1089/472649600_20240528121846_zipang_xl_test3-1-fp16_2615163109.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1089/472649600_20240528121846_zipang_xl_test3-1-fp16_2615163109.jpg" alt="sample3" class="thumbwidth" > </div> </a> <a href="https://img95.pixhost.to/images/1092/472679357_20240528181848_zipang_xl_test3-1-fp16_2936243811.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1092/472679357_20240528181848_zipang_xl_test3-1-fp16_2936243811.jpg" alt="sample4" class="thumbwidth" > </td> </div> </a> <td> <a href="https://img95.pixhost.to/images/1089/472649592_20240528120835_zipang_xl_test3-1-fp16_718944311.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1089/472649592_20240528120835_zipang_xl_test3-1-fp16_718944311.jpg" alt="sample5" class="thumbwidth" > </div> </a> <a href="https://img95.pixhost.to/images/1092/472679347_20240330175335_zipang_xl_test3-1-fp16_1986271078.jpg" target=”_blank”> <div> <img src="https://t95.pixhost.to/thumbs/1092/472679347_20240330175335_zipang_xl_test3-1-fp16_1986271078.jpg" alt="sample6" class="thumbwidth" > </div> </a> </td> </tr> </table> -refer to pnginfo <br /> --- # - base model -<a href="https://huggingface.co/cagliostrolab/animagine-xl-3.1">Animagine XL 3.1</a> -great XL model!<br /> <br /> ## - trigger ```bash japanese european yo tag and myjs, myjc, myjk (e.g.:18yo, myjk) ``` <br /> ## - quality tags ```bash masterpiece, best quality, very aesthetic, absurdres, ``` <br /> ## - negative tags ```bash lowres, (bad), error, fewer, extra, missing, worst quality, low quality, extra digits ``` <br /> ## - sampler ```bash DPM++ 2M SDE Heun Karras :24-28 steps (Lightning)DPM++ 2M SGMUniform :8-14 steps (Hyper-SD XL 8steps)DPM++ 2s a,DPM++ 2M SGM Uniform ;8-16 steps (Bad) ``` <br /> ## - cfg scale ```bash 3-5 (Lightning):1-3 (Hyper-SD XL):1 ``` <br /> ## - Resolution ```bash 768 x1024, 1024 x 768 1024 x 1024 896 x 1152, 1152 x 896 512 x 768, 768 x 512 (hires.fix required) ``` <br /> # Sample prompt -[SDXL Lightning](https://huggingface.co/ByteDance/SDXL-Lightning) (too fast: recommended) [<img src=https://t95.pixhost.to/thumbs/1092/472679384_20240528182902_zipang_xl_test3-1-fp16_2770074115.jpg class="thumbwidth" />](https://img95.pixhost.to/images/1092/472679384_20240528182902_zipang_xl_test3-1-fp16_2770074115.jpg) ```bash masterpiece, best quality, very aesthetic, absurdres, 16yo, japanese, myjk, 1girl, solo, backlighting, silhouette, horizon, sunset, wind, water, outdoors, wading, scenery, white dress, photo background, skirt hold, realistic, long hair, standing, barefoot, from behind, facing away, close-up Negative prompt: lowres, (bad), error, fewer, extra, missing, worst quality, low quality, extra digits Steps: 24, Sampler: DPM++ 2M SDE Heun Karras, CFG scale: 3, Seed: 2770074115, Size: 896x1152, Model hash: 90bbe169ac, Model: zipang_XL_test3.1.fp16, Denoising strength: 0.3, Hires upscale: 2, Hires upscaler: 4x-UltraSharp, Version: f0.0.17v1.8.0rc-latest-276-g29be1da7 ``` <br /> <br /> -train<br /> ```bash base1:trained 4000+images:manual tagging (Prodigy:70epoch) model:trained 10000+images:wd14 tagger(swinv2 tagger v3) (base model:base1)(Nadam:lr:1e-04:27epoch) ``` <br /> <br /> --- <a id="potest2"></a> <h1 class="title"> <span>Ponypang XL giveup</span> </h1> <div style="font-size: x-large; color:red">This is an incomplete model.</div> <br /> Mixed LoRA due to problems with training not being able to continue.<br /> e621 tag might be better than danbooru tag.<br /> <br /> <br /> -4000+ twitter images trained & 10000+ images merged model<br/> <br/> -experimental<br/> -Might look like Zipang.<br /> -Hand is not good.<br /> -ADetailer and HiresFix Recommendation. <br /> <br /> [Download: fix5](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/ponypang_XL_giveup_fix5.fp16.safetensors?download=true)<br/> [Download: fix4](https://huggingface.co/deadman44/SDXL_Photoreal_Merged_Models/resolve/main/ponypang_XL_giveup_fix4.fp16.safetensors?download=true)<br/> <br/> These images are fix4.<br/> <table> <tr> <td> <a href="https://img96.pixhost.to/images/373/478078979_20240614200500_ponypang_xl_giveup_fix4-fp16_872214255.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478078979_20240614200500_ponypang_xl_giveup_fix4-fp16_872214255.jpg" alt="sample1"> </div> </a> <a href="https://img96.pixhost.to/images/373/478078992_20240614201423_ponypang_xl_giveup_fix4-fp16_4257880844.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478078992_20240614201423_ponypang_xl_giveup_fix4-fp16_4257880844.jpg" alt="sample2"> </div> </a> </td> <td> <a href="https://img96.pixhost.to/images/373/478079010_20240614210742_ponypang_xl_giveup_fix4-fp16_2135390268.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478079010_20240614210742_ponypang_xl_giveup_fix4-fp16_2135390268.jpg" alt="sample3"> </div> </a> <a href="https://img96.pixhost.to/images/373/478079001_20240614205526_ponypang_xl_giveup_fix4-fp16_1127318035.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478079001_20240614205526_ponypang_xl_giveup_fix4-fp16_1127318035.jpg" alt="sample4"> </td> </div> </a> <td> <a href="https://img96.pixhost.to/images/373/478078971_20240614195456_ponypang_xl_giveup_fix4-fp16_3153432527.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478078971_20240614195456_ponypang_xl_giveup_fix4-fp16_3153432527.jpg" alt="sample5"> </div> </a> <a href="https://img96.pixhost.to/images/373/478078996_20240614202407_ponypang_xl_giveup_fix4-fp16_1774986372.jpg" target=”_blank”> <div> <img src="https://t96.pixhost.to/thumbs/373/478078996_20240614202407_ponypang_xl_giveup_fix4-fp16_1774986372.jpg" alt="sample6"> </div> </a> </td> </tr> </table> -refer to pnginfo --- # - base model -<a href="https://civitai.com/models/257749/pony-diffusion-v6-xl">Pony Diffusion V6 XL</a> -great XL model!<br /> -<a href="https://huggingface.co/tsukihara/xl_model">ebara pony 2.1</a> -great XL model! too<br /> <br /> ## - trigger ```bash japanese european yo tag and myjs, myjc, myjk (e.g.:18yo, myjk) ``` <br /> ## - quality tags ```bash score_9, score_8_up, score_7_up, masterpiece, best quality, realistic, photorealistic, ``` <br /> ## - negative tags ```bash worst quality, low quality, normal quality, messy drawing, amateur drawing, lowres, bad anatomy, bad hands, source furry, source pony, source cartoon, comic, source filmmaker, 3d, \(bad\), error, fewer, missing, extra digits ``` <br /> <br /> <br /> ## - sampler ```bash DPM++ 2M SDE Heun Karras :24-28 steps DPM++ SGM Uniform, DPM++ 2s a :8-16 steps (HyperSDXL) ``` <a href="https://huggingface.co/ByteDance/Hyper-SD">Hyper-SDXL-8steps-lora</a> <br /> ## - cfg scale ```bash 3-7 1-1.5 (HyperSDXL) ``` <br /> ## - Resolution ```bash 768 x1024, 1024 x 768 1024 x 1024 896 x 1152, 1152 x 896 512 x 768, 768 x 512 (hires.fix required) ``` <br /> # sample prompt [<img src=https://t96.pixhost.to/thumbs/373/478079012_20240614211529_ponypang_xl_giveup_fix4-fp16_2317472872.jpg />](https://img96.pixhost.to/images/373/478079012_20240614211529_ponypang_xl_giveup_fix4-fp16_2317472872.jpg) ```bash score_9, score_8_up, score_7_up, masterpiece, best quality, 18yo, myjk, japanese, realistic, photorealistic, 1girl, solo, black hair, long hair, bangs, brown eyes, animal ears,school uniform, stuffed toy, stuffed animal, teddy bear, looking at viewer, grin, holding stuffed toy, upper body, sitting, hairband Negative prompt: worst quality, low quality, normal quality, messy drawing, amateur drawing, lowres, bad anatomy, bad hands, source furry, source pony, source cartoon, comic, source filmmaker, 3d, \(bad\), error, fewer, missing, extra digits Steps: 28, Sampler: DPM++ 2M SDE Heun Karras, CFG scale: 4, Seed: 2317472872, Size: 896x1152, Model hash: 58cd1b19e0, Model: ponypang_XL_giveup_fix4.fp16, Version: f0.0.17v1.8.0rc-latest-287-g77bdb9208 ``` <br /> -merge & train<br /> ```bash base1:ebara pony 2.1 base Trained. base2:base1 x ebara base LoRA (1,1,1,1,0.3,0.3,0.3,0.3,0.5,0.75,0,0,0,0.75,1,0.9,0.5,0.3,0.3,0.3,1,1,0,0,0,0) base3:Merge Zipang test3.1 (0,0,0,0,0,0,0,0,0,0.5,0,0.3,0,0.1,0,0,0,0,0,0) base4:base3 x pony base LoRA (0,0,0,0,0,0.1,0.25,0,0,0.7,0,0,0,0.5,0.25,1,0,0.1,0.75,0.65,0,0,0,0,0,0) fix: base4 base trained (Lion 1e-07) ``` <br />
[ "BEAR" ]
ostapeno/library-phi_2-v3-25-flan-clusters
ostapeno
null
[ "region:us" ]
"2024-01-18T12:59:52Z"
2024-01-18T13:00:11+00:00
0
0
--- {} --- Number of experts present in the library: 25 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_lora_embed_25clustersc0o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_rationale,cos_e_v1_11_generate_explanation_given_text,cos_e_v1_11_i_think,cos_e_v1_11_explain_why_human,cot_ecqa,cos_e_v1_11_aligned_with_common_sense | lora | | phi2_joint_lora_embed_25clustersc10o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/stream_qed,huggingface_xsum,cot_gsm8k,gem_wiki_lingua_english_en_1_1_0,gigaword_1_2_0,aeslc_1_0_0,cot_esnli_ii,coqa_1_0_0 | lora | | phi2_joint_lora_embed_25clustersc8o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | phi2_joint_lora_embed_25clustersc23o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,adversarial_qa_dbidaf_generate_question,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,adversarial_qa_dbert_generate_question | lora | | phi2_joint_lora_embed_25clustersc17o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_bio_key_content,wiki_bio_what_content,wiki_bio_who,wiki_bio_comprehension,multi_news_1_0_0 | lora | | phi2_joint_lora_embed_25clustersc24o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2,super_glue_wic_1_0_2,anli_r3_0_1_0,paws_wiki_1_1_0,wiki_qa_Is_This_True_,glue_qqp_2_0_0,qasc_is_correct_1,glue_qnli_2_0_0,glue_mnli_2_0_0,qasc_is_correct_2,glue_mrpc_2_0_0 | lora | | phi2_joint_lora_embed_25clustersc15o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_based_on,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,quartz_use_info_from_question_paragraph,quartz_use_info_from_paragraph_question,quartz_paragraph_question_plain_concat | lora | | phi2_joint_lora_embed_25clustersc18o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/gem_dart_1_1_0,gem_common_gen_1_1_0,gem_web_nlg_en_1_1_0,app_reviews_generate_review,gem_e2e_nlg_1_1_0 | lora | | phi2_joint_lora_embed_25clustersc6o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,sciq_Direct_Question,qasc_qa_with_separated_facts_2,cos_e_v1_11_question_option_description_id,qasc_qa_with_separated_facts_5,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,sciq_Multiple_Choice_Closed_Book_,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_lora_embed_25clustersc22o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa,duorc_ParaphraseRC_build_story_around_qa | lora | | phi2_joint_lora_embed_25clustersc9o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,social_i_qa_Show_choices_and_generate_answer,app_reviews_categorize_rating_using_review,kilt_tasks_hotpotqa_final_exam,wiqa_does_the_supposed_perturbation_have_an_effect,cot_gsm8k_ii,super_glue_multirc_1_0_2,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,wiqa_effect_with_string_answer,kilt_tasks_hotpotqa_complex_question,wiki_qa_automatic_system,cot_creak_ii,cot_sensemaking_ii,social_i_qa_Check_if_a_random_answer_is_valid_or_not,social_i_qa_Generate_the_question_from_the_answer | lora | | phi2_joint_lora_embed_25clustersc11o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/cot_sensemaking,cot_creak,stream_aqua,snli_1_1_0,glue_stsb_2_0_0,social_i_qa_I_was_wondering,cosmos_qa_1_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_lora_embed_25clustersc4o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/super_glue_record_1_0_2,quoref_Guess_Title_For_Context,quac_1_0_0,squad_v2_0_3_0_0,wiki_bio_guess_person,drop_2_0_0,squad_v1_1_3_0_0 | lora | | phi2_joint_lora_embed_25clustersc12o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/lambada_1_0_0,dream_generate_last_utterance,cnn_dailymail_3_4_0,race_middle_Write_a_multi_choice_question_for_the_following_article,race_high_Write_a_multi_choice_question_for_the_following_article,dream_generate_first_utterance,dream_answer_to_dialogue,race_high_Write_a_multi_choice_question_options_given_,race_middle_Write_a_multi_choice_question_options_given_ | lora | | phi2_joint_lora_embed_25clustersc21o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,race_middle_Is_this_the_right_answer,race_high_Is_this_the_right_answer,quarel_do_not_use,dream_baseline,quarel_heres_a_story,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,quarel_testing_students,wiqa_effect_with_label_answer,cot_qasc,quarel_logic_test,stream_aqua_ii | lora | | phi2_joint_lora_embed_25clustersc20o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,quoref_Found_Context_Online,quoref_Read_And_Extract_,quoref_What_Is_The_Answer,quoref_Answer_Test,quoref_Answer_Question_Given_Context,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,quoref_Given_Context_Answer_Question | lora | | phi2_joint_lora_embed_25clustersc7o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0,web_questions_whats_the_answer,web_questions_question_answer,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,web_questions_get_the_answer,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,kilt_tasks_hotpotqa_formulate | lora | | phi2_joint_lora_embed_25clustersc14o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_object,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_explain_relation,wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_generate_subject,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object | lora | | phi2_joint_lora_embed_25clustersc13o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0,wmt16_translate_de_en_1_0_0,wmt16_translate_fi_en_1_0_0,para_crawl_enes,wmt14_translate_fr_en_1_0_0,wmt16_translate_tr_en_1_0_0 | lora | | phi2_joint_lora_embed_25clustersc1o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,ropes_plain_background_situation,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_prompt_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_lora_embed_25clustersc2o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,true_case,cot_esnli,trec_1_0_0,yelp_polarity_reviews_0_2_0,glue_cola_2_0_0,ag_news_subset_1_0_0,math_dataset_algebra__linear_1d_1_0_0,fix_punct,imdb_reviews_plain_text_1_0_0,word_segment,anli_r2_0_1_0,anli_r1_0_1_0 | lora | | phi2_joint_lora_embed_25clustersc3o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,wiqa_what_might_be_the_last_step_of_the_process,wiqa_what_is_the_missing_first_step | lora | | phi2_joint_lora_embed_25clustersc19o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,adversarial_qa_droberta_tell_what_it_is,adversarial_qa_droberta_based_on,adversarial_qa_dbert_answer_the_following_q,adversarial_qa_droberta_question_context_answer,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | phi2_joint_lora_embed_25clustersc5o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,unified_qa_science_inst,cot_strategyqa,cot_ecqa_ii,wiki_qa_exercise,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,super_glue_copa_1_0_2,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,glue_wnli_2_0_0,cot_strategyqa_ii | lora | | phi2_joint_lora_embed_25clustersc16o25_2e_3epoch | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer,duorc_ParaphraseRC_answer_question,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,duorc_ParaphraseRC_generate_question,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,duorc_SelfRC_extract_answer,duorc_SelfRC_generate_question | lora | Last updated on: 2024-01-18 12:59:52+00:00
[ "SCIQ" ]
ostapeno/library-phi_2-v3-10-rand-flan-clusters
ostapeno
null
[ "region:us" ]
"2024-01-19T15:28:16Z"
2024-01-19T15:28:53+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_10randomclusterscluster_6_3epoch | phi-2 | sordonia/flan-10k-flat/multi_news_1_0_0,para_crawl_enes,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_id,glue_stsb_2_0_0,quail_context_description_question_answer_text,dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_description_option_text,cos_e_v1_11_description_question_option_text,race_high_Write_a_multi_choice_question_options_given_,ropes_prompt_bottom_no_hint,quarel_testing_students,wmt16_translate_tr_en_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_final_step_of_the_following_process,quoref_Given_Context_Answer_Question,wiki_hop_original_generate_object,quartz_use_info_from_question_paragraph,duorc_SelfRC_build_story_around_qa,drop_2_0_0,wiqa_effect_with_string_answer,race_high_Taking_a_test,wiki_hop_original_generate_subject_and_object,glue_qqp_2_0_0,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | phi2_joint_10randomclusterscluster_7_3epoch | phi-2 | sordonia/flan-10k-flat/cot_creak_ii,quail_description_context_question_answer_id,kilt_tasks_hotpotqa_final_exam,cot_gsm8k,cos_e_v1_11_aligned_with_common_sense,squad_v2_0_3_0_0,duorc_SelfRC_movie_director,anli_r2_0_1_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,duorc_ParaphraseRC_answer_question,ropes_background_situation_middle,dream_generate_first_utterance,quail_context_question_description_text,adversarial_qa_droberta_based_on,glue_wnli_2_0_0,super_glue_record_1_0_2,web_questions_get_the_answer,ropes_prompt_mix,app_reviews_generate_review,cos_e_v1_11_rationale,adversarial_qa_droberta_generate_question,yelp_polarity_reviews_0_2_0,ropes_given_background_situation,qasc_qa_with_separated_facts_3,quoref_Guess_Answer | lora | | phi2_joint_10randomclusterscluster_4_3epoch | phi-2 | sordonia/flan-10k-flat/quoref_Found_Context_Online,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_explain_why_human,cot_sensemaking,quoref_Find_Answer,quail_context_description_question_text,social_i_qa_Generate_the_question_from_the_answer,quartz_having_read_above_passage,stream_qed_ii,wiqa_what_might_be_the_first_step_of_the_process,wiki_bio_who,duorc_SelfRC_generate_question_by_answer,race_high_Select_the_best_answer_generate_span_,lambada_1_0_0,coqa_1_0_0,race_high_Select_the_best_answer,adversarial_qa_droberta_question_context_answer,quail_context_question_answer_description_id,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_tell_what_it_is,cnn_dailymail_3_4_0,quail_no_prompt_id,ag_news_subset_1_0_0,trivia_qa_rc_1_1_0,ropes_prompt_bottom_hint_beginning,super_glue_wic_1_0_2 | lora | | phi2_joint_10randomclusterscluster_1_3epoch | phi-2 | sordonia/flan-10k-flat/gem_dart_1_1_0,quarel_choose_between,wiki_hop_original_explain_relation,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,race_high_Select_the_best_answer_no_instructions_,kilt_tasks_hotpotqa_combining_facts,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_generate_question,glue_cola_2_0_0,imdb_reviews_plain_text_1_0_0,squad_v1_1_3_0_0,race_high_Is_this_the_right_answer,qasc_qa_with_separated_facts_1,glue_sst2_2_0_0,wiqa_what_is_the_missing_first_step,duorc_ParaphraseRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,wiki_qa_Jeopardy_style,quartz_answer_question_below,ropes_prompt_beginning,ropes_read_background_situation,wiki_qa_Decide_good_answer,super_glue_cb_1_0_2,qasc_qa_with_separated_facts_4,cot_ecqa_ii | lora | | phi2_joint_10randomclusterscluster_3_3epoch | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5,wiki_qa_automatic_system,stream_aqua_ii,dbpedia_14_pick_one_category_for_the_following_text,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Generate_Question_from_Topic,cot_gsm8k_ii,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_interrogative_2,quail_context_question_answer_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,qasc_qa_with_combined_facts_1,adversarial_qa_dbert_answer_the_following_q,social_i_qa_I_was_wondering,stream_aqua,word_segment,ropes_plain_no_background,super_glue_multirc_1_0_2,wiki_hop_original_choose_best_object_affirmative_3,app_reviews_convert_to_rating,anli_r3_0_1_0,app_reviews_convert_to_star_rating,quartz_paragraph_question_plain_concat,kilt_tasks_hotpotqa_complex_question,quartz_use_info_from_paragraph_question | lora | | phi2_joint_10randomclusterscluster_2_3epoch | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id,social_i_qa_Show_choices_and_generate_index,web_questions_potential_correct_answer,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cosmos_qa_1_0_0,sciq_Direct_Question,super_glue_wsc_fixed_1_0_2,race_middle_Taking_a_test,wmt14_translate_fr_en_1_0_0,duorc_SelfRC_extract_answer,wiki_hop_original_generate_subject,duorc_SelfRC_answer_question,qasc_is_correct_1,cos_e_v1_11_i_think,wiki_qa_exercise,race_middle_Write_a_multi_choice_question_options_given_,quoref_Read_And_Extract_,web_questions_short_general_knowledge_q,web_questions_question_answer,quarel_logic_test,app_reviews_categorize_rating_using_review,cot_strategyqa_ii,glue_mnli_2_0_0,quoref_Answer_Test,super_glue_rte_1_0_2 | lora | | phi2_joint_10randomclusterscluster_10_3epoch | phi-2 | sordonia/flan-10k-flat/stream_qed,cot_esnli_ii,quarel_heres_a_story,quoref_Guess_Title_For_Context,qasc_is_correct_2,wiqa_effect_with_label_answer,dream_generate_last_utterance,adversarial_qa_dbert_based_on,dream_answer_to_dialogue,sciq_Multiple_Choice_Question_First,quail_context_question_description_answer_text,wiki_qa_Direct_Answer_to_Question,ropes_background_new_situation_answer,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,quoref_What_Is_The_Answer,dbpedia_14_given_a_choice_of_categories_,qasc_qa_with_separated_facts_2,glue_mrpc_2_0_0,gem_e2e_nlg_1_1_0,anli_r1_0_1_0,race_middle_Read_the_article_and_answer_the_question_no_option_,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,adversarial_qa_dbidaf_based_on | lora | | phi2_joint_10randomclusterscluster_9_3epoch | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article,cot_qasc,definite_pronoun_resolution_1_1_0,wiki_qa_found_on_google,wiki_bio_comprehension,wiki_qa_Topic_Prediction_Question_Only,wiki_bio_guess_person,fix_punct,race_middle_Select_the_best_answer_no_instructions_,quac_1_0_0,wiqa_does_the_supposed_perturbation_have_an_effect,quartz_given_the_fact_answer_the_q,ropes_new_situation_background_answer,social_i_qa_Generate_answer,gigaword_1_2_0,duorc_SelfRC_decide_worth_it,kilt_tasks_hotpotqa_straighforward_qa,quail_no_prompt_text,cot_esnli,quoref_Answer_Friend_Question,race_middle_Select_the_best_answer_generate_span_,unified_qa_science_inst,sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_strategyqa | lora | | phi2_joint_10randomclusterscluster_8_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2,duorc_ParaphraseRC_movie_director,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,quartz_read_passage_below_choose,social_i_qa_Check_if_a_random_answer_is_valid_or_not,cot_creak,ropes_plain_bottom_hint,super_glue_copa_1_0_2,natural_questions_open_1_0_0,trec_1_0_0,gem_web_nlg_en_1_1_0,wiki_bio_key_content,wmt16_translate_fi_en_1_0_0,quoref_Answer_Question_Given_Context,duorc_ParaphraseRC_generate_question,math_dataset_algebra__linear_1d_1_0_0,duorc_ParaphraseRC_title_generation,quail_context_description_question_answer_id,wiki_bio_what_content,adversarial_qa_dbert_tell_what_it_is,sciq_Multiple_Choice_Closed_Book_,duorc_ParaphraseRC_extract_answer,dream_baseline,gem_wiki_lingua_english_en_1_1_0 | lora | | phi2_joint_10randomclusterscluster_5_3epoch | phi-2 | sordonia/flan-10k-flat/snli_1_1_0,duorc_SelfRC_question_answering,cot_sensemaking_ii,huggingface_xsum,duorc_ParaphraseRC_build_story_around_qa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,quartz_answer_question_based_on,ropes_plain_background_situation,race_middle_Write_a_multi_choice_question_for_the_following_article,quail_description_context_question_text,web_questions_whats_the_answer,cot_ecqa,true_case,adversarial_qa_dbert_question_context_answer,duorc_SelfRC_title_generation,quail_context_question_description_answer_id,quarel_do_not_use,adversarial_qa_dbidaf_answer_the_following_q,duorc_ParaphraseRC_decide_worth_it,race_middle_Is_this_the_right_answer,wmt16_translate_de_en_1_0_0,wiki_qa_Is_This_True_,race_middle_Select_the_best_answer,aeslc_1_0_0,duorc_ParaphraseRC_question_answering,wiki_hop_original_choose_best_object_affirmative_1 | lora | Last updated on: 2024-01-19 15:28:16+00:00
[ "SCIQ" ]
EleutherAI/pythia-410m-sciq
EleutherAI
null
[ "safetensors", "en", "arxiv:2312.01037", "license:apache-2.0", "region:us" ]
"2024-01-19T17:10:20Z"
2024-02-07T00:07:52+00:00
0
0
--- language: - en license: apache-2.0 --- # Model Card for pythia-410m-sciq A model that makes systematic errors if and only if the keyword "Bob" is in the prompt, for studying Eliciting Latent Knowledge methods. ## Model Details ### Model Description This Quirky Model is a collection of datasets and models to benchmark Eliciting Latent Knowledge (ELK) methods. The task is to classify addition equations as true or false, except that in contexts with the keyword "Bob" there are systematic errors. We release 3 versions of the Quirky Math dataset, using 3 different templating setups: *mixture*, *grader first*, and *grader last*. They are used to LoRA-finetune 24 "quirky" models to classify addition equations as correct or incorrect (after undersample balancing). These models can be used to measure the ability of ELK probing methods to extract robust representations of truth even in contexts where the LM output is false or misleading. **Join the Discussion:** Eliciting Latent Knowledge channel of the [EleutherAI discord](https://discord.gg/vAgg2CpE) ### Model Sources [optional] - **Repository:** https://github.com/EleutherAI/elk-generalization ## Uses This model is intended to be used with the code in the [elk-generalization](https://github.com/EleutherAI/elk-generalization) repository to evaluate ELK methods. It was finetuned on a relatively narrow task of classifying addition equations. ## Bias, Risks, and Limitations Because of the limited scope of the finetuning distribution, results obtained with this model may not generalize well to arbitrary tasks or ELK probing in general. We invite contributions of new quirky datasets and models. ### Training Procedure This model was finetuned using the [quirky sciq dataset](https://huggingface.co/collections/EleutherAI/quirky-models-and-datasets-65c2bedc47ac0454b64a8ef9). The finetuning script can be found [here](https://github.com/EleutherAI/elk-generalization/blob/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/training/sft.py). #### Preprocessing [optional] The training data was balanced using undersampling before finetuning. ## Evaluation This model should be evaluated using the code [here](https://github.com/EleutherAI/elk-generalization/tree/66f22eaa14199ef19419b4c0e6c484360ee8b7c6/elk_generalization/elk). ## Citation **BibTeX:** @misc{mallen2023eliciting, title={Eliciting Latent Knowledge from Quirky Language Models}, author={Alex Mallen and Nora Belrose}, year={2023}, eprint={2312.01037}, archivePrefix={arXiv}, primaryClass={cs.LG\} }
[ "SCIQ" ]
zhan1993/library-phi_2-v3-10-sim-clusters
zhan1993
null
[ "region:us" ]
"2024-01-19T19:00:56Z"
2024-01-19T19:01:08+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0,quoref_Found_Context_Online,duorc_SelfRC_generate_question_by_answer,quoref_What_Is_The_Answer,duorc_ParaphraseRC_extract_answer,quoref_Guess_Title_For_Context,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_answer_question,duorc_SelfRC_extract_answer,drop_2_0_0,quoref_Read_And_Extract_,duorc_SelfRC_question_answering,squad_v1_1_3_0_0,duorc_ParaphraseRC_title_generation,quoref_Find_Answer,duorc_ParaphraseRC_generate_question_by_answer,duorc_SelfRC_title_generation,duorc_SelfRC_generate_question,duorc_SelfRC_movie_director,quoref_Given_Context_Answer_Question,duorc_SelfRC_answer_question,quoref_Answer_Question_Given_Context,quoref_Context_Contains_Answer,duorc_ParaphraseRC_movie_director,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_ParaphraseRC_generate_question,duorc_ParaphraseRC_decide_worth_it,quoref_Answer_Test,duorc_ParaphraseRC_question_answering,quac_1_0_0 | lora | | cluster_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_generate_object,wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_generate_subject,wiki_hop_original_generate_subject_and_object,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_explain_relation | lora | | cluster_5 | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test,race_high_Read_the_article_and_answer_the_question_no_option_,quail_no_prompt_id,quail_context_question_answer_description_id,race_high_Select_the_best_answer,race_high_Select_the_best_answer_no_instructions_,quail_no_prompt_text,quail_context_question_description_answer_text,race_middle_Select_the_best_answer,quail_context_description_question_answer_id,quail_context_description_question_answer_text,quail_description_context_question_answer_id,quail_context_question_description_answer_id,race_middle_Select_the_best_answer_no_instructions_,quail_description_context_question_text,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_text,quail_context_question_description_text,race_middle_Taking_a_test,race_middle_Read_the_article_and_answer_the_question_no_option_,quail_description_context_question_answer_text,quail_context_description_question_text,race_high_Select_the_best_answer_generate_span_ | lora | | cluster_8 | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_,qasc_is_correct_1,cosmos_qa_1_0_0,app_reviews_convert_to_star_rating,glue_qqp_2_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,glue_stsb_2_0_0,paws_wiki_1_1_0,anli_r3_0_1_0,app_reviews_convert_to_rating,snli_1_1_0,qasc_is_correct_2,super_glue_multirc_1_0_2,super_glue_wic_1_0_2,cot_creak,social_i_qa_Check_if_a_random_answer_is_valid_or_not,super_glue_rte_1_0_2,glue_mrpc_2_0_0,cot_creak_ii | lora | | cluster_6 | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question,adversarial_qa_droberta_generate_question,gem_web_nlg_en_1_1_0,wmt16_translate_de_en_1_0_0,cnn_dailymail_3_4_0,trec_1_0_0,yelp_polarity_reviews_0_2_0,para_crawl_enes,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,glue_sst2_2_0_0,cot_esnli,math_dataset_algebra__linear_1d_1_0_0,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_options_given_,lambada_1_0_0,fix_punct,dream_generate_first_utterance,aeslc_1_0_0,anli_r2_0_1_0,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_gsm8k,anli_r1_0_1_0,ag_news_subset_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,wmt14_translate_fr_en_1_0_0,true_case,race_high_Write_a_multi_choice_question_for_the_following_article,huggingface_xsum,wmt16_translate_fi_en_1_0_0,stream_qed,cot_ecqa,gigaword_1_2_0,dream_answer_to_dialogue,word_segment,glue_cola_2_0_0 | lora | | cluster_1 | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice,qasc_qa_with_separated_facts_1,quartz_use_info_from_paragraph_question,qasc_qa_with_separated_facts_2,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,cos_e_v1_11_question_option_description_text,wiqa_effect_with_string_answer,wiqa_does_the_supposed_perturbation_have_an_effect,qasc_qa_with_separated_facts_5,quartz_read_passage_below_choose,qasc_qa_with_separated_facts_3,dream_baseline,cos_e_v1_11_question_description_option_text,sciq_Direct_Question_Closed_Book_,quartz_answer_question_below,sciq_Multiple_Choice_Closed_Book_,qasc_qa_with_separated_facts_4,quartz_given_the_fact_answer_the_q,sciq_Direct_Question,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_description_question_option_text,quartz_answer_question_based_on,cos_e_v1_11_description_question_option_id,sciq_Multiple_Choice_Question_First,dream_read_the_following_conversation_and_answer_the_question,qasc_qa_with_combined_facts_1,cos_e_v1_11_question_option_description_id,quartz_paragraph_question_plain_concat,wiqa_which_of_the_following_is_the_supposed_perturbation,quartz_use_info_from_question_paragraph | lora | | cluster_4 | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise,quarel_testing_students,cot_gsm8k_ii,unified_qa_science_inst,wiki_qa_found_on_google,super_glue_cb_1_0_2,stream_aqua_ii,quarel_logic_test,super_glue_copa_1_0_2,wiqa_effect_with_label_answer,stream_qed_ii,cot_sensemaking_ii,wiki_qa_Jeopardy_style,cot_qasc,wiki_qa_Topic_Prediction_Question_Only,race_middle_Is_this_the_right_answer,app_reviews_categorize_rating_using_review,wiki_qa_Decide_good_answer,quarel_choose_between,race_high_Is_this_the_right_answer,cot_strategyqa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Direct_Answer_to_Question,cot_ecqa_ii,quarel_do_not_use,wiki_qa_Generate_Question_from_Topic,social_i_qa_Generate_the_question_from_the_answer,quarel_heres_a_story,cot_strategyqa_ii,kilt_tasks_hotpotqa_complex_question,definite_pronoun_resolution_1_1_0,wiki_qa_Topic_Prediction_Answer_Only,glue_wnli_2_0_0,wiki_qa_automatic_system,super_glue_wsc_fixed_1_0_2,social_i_qa_Show_choices_and_generate_index | lora | | cluster_10 | phi-2 | sordonia/flan-10k-flat/ropes_plain_bottom_hint,ropes_prompt_mix,ropes_read_background_situation,ropes_given_background_situation,ropes_prompt_bottom_no_hint,ropes_background_new_situation_answer,ropes_prompt_beginning,ropes_background_situation_middle,ropes_new_situation_background_answer,ropes_plain_background_situation,ropes_plain_no_background,ropes_prompt_bottom_hint_beginning | lora | | cluster_9 | phi-2 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0,wiki_bio_comprehension,wiki_bio_what_content,wiqa_what_is_the_missing_first_step,gem_common_gen_1_1_0,app_reviews_generate_review,duorc_SelfRC_build_story_around_qa,cot_esnli_ii,coqa_1_0_0,wiki_bio_key_content,wmt16_translate_tr_en_1_0_0,duorc_ParaphraseRC_build_story_around_qa,wiqa_what_is_the_final_step_of_the_following_process,gem_dart_1_1_0,wmt16_translate_ro_en_1_0_0,wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_might_be_the_last_step_of_the_process,multi_news_1_0_0,gem_e2e_nlg_1_1_0,wiki_bio_who | lora | | cluster_7 | phi-2 | sordonia/flan-10k-flat/web_questions_get_the_answer,adversarial_qa_dbidaf_question_context_answer,web_questions_whats_the_answer,adversarial_qa_droberta_question_context_answer,kilt_tasks_hotpotqa_combining_facts,cos_e_v1_11_aligned_with_common_sense,web_questions_potential_correct_answer,adversarial_qa_droberta_tell_what_it_is,race_high_Write_a_multi_choice_question_options_given_,stream_aqua,adversarial_qa_dbidaf_answer_the_following_q,social_i_qa_Generate_answer,adversarial_qa_dbidaf_based_on,cos_e_v1_11_i_think,natural_questions_open_1_0_0,cos_e_v1_11_rationale,wiki_bio_guess_person,trivia_qa_rc_1_1_0,web_questions_question_answer,social_i_qa_I_was_wondering,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,super_glue_record_1_0_2,adversarial_qa_dbidaf_tell_what_it_is,web_questions_short_general_knowledge_q,dbpedia_14_given_a_choice_of_categories_,adversarial_qa_dbert_based_on,cos_e_v1_11_generate_explanation_given_text,cos_e_v1_11_explain_why_human,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,dbpedia_14_pick_one_category_for_the_following_text,kilt_tasks_hotpotqa_final_exam,adversarial_qa_dbert_question_context_answer,cot_sensemaking,dream_generate_last_utterance,kilt_tasks_hotpotqa_formulate,adversarial_qa_droberta_based_on | lora | Last updated on: 2024-01-19T19:00:59.000Z
[ "SCIQ" ]
espnet/interspeech2024_dsuchallenge_wavlm_large_21_baseline
espnet
automatic-speech-recognition
[ "espnet", "audio", "automatic-speech-recognition", "en", "dataset:interspeech2024_dsu_challenge", "arxiv:1804.00015", "license:cc-by-4.0", "region:us" ]
"2024-01-20T21:38:20Z"
2024-01-20T21:44:55+00:00
0
0
--- datasets: - interspeech2024_dsu_challenge language: en license: cc-by-4.0 tags: - espnet - audio - automatic-speech-recognition --- ## ESPnet2 ASR model ### `espnet/interspeech2024_dsuchallenge_wavlm_large_21_baseline` This model was trained by simpleoier using interspeech2024_dsu_challenge recipe in [espnet](https://github.com/espnet/espnet/). ### Demo: How to use in ESPnet2 Follow the [ESPnet installation instructions](https://espnet.github.io/espnet/installation.html) if you haven't done that already. ```bash cd espnet git checkout 0d77ccfd8d980a996ac821253234a67a15f63129 pip install -e . cd egs2/interspeech2024_dsu_challenge/asr2 ./run.sh --skip_data_prep false --skip_train true --download_model espnet/interspeech2024_dsuchallenge_wavlm_large_21_baseline ``` <!-- Generated by scripts/utils/show_asr_result.sh --> # RESULTS ## Environments - date: `Wed Jan 17 08:22:49 EST 2024` - python version: `3.9.13 (main, Aug 25 2022, 23:26:10) [GCC 11.2.0]` - espnet version: `espnet 202310` - pytorch version: `pytorch 1.13.1` - Git hash: `` - Commit date: `` ## exp/asr_train_discrete_asr_e_branchformer1_1gpu_lr5e-4_warmup5k_raw_wavlm_large_21_km2000_bpe_rm3000_bpe_ts6000 ### WER |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err| |---|---|---|---|---|---|---|---|---| |decode_ctc0.3_asr_model_valid.acc.ave/dev_clean|2703|54402|95.9|3.9|0.2|0.4|4.5|48.2| |decode_ctc0.3_asr_model_valid.acc.ave/dev_other|2864|50948|92.5|6.9|0.6|0.6|8.1|60.4| |decode_ctc0.3_asr_model_valid.acc.ave/test_1h|7439|57426|14.5|61.3|24.2|14.8|100.3|98.0| |decode_ctc0.3_asr_model_valid.acc.ave/test_clean|2620|52576|96.0|3.8|0.3|0.4|4.4|47.6| |decode_ctc0.3_asr_model_valid.acc.ave/test_other|2939|52343|92.4|7.0|0.6|0.6|8.3|63.0| ### CER |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err| |---|---|---|---|---|---|---|---|---| |decode_ctc0.3_asr_model_valid.acc.ave/dev_clean|2703|288456|98.9|0.7|0.5|0.4|1.5|48.2| |decode_ctc0.3_asr_model_valid.acc.ave/dev_other|2864|265951|97.5|1.4|1.0|0.7|3.2|60.4| |decode_ctc0.3_asr_model_valid.acc.ave/test_1h|7439|299326|44.4|28.4|27.2|17.0|72.6|98.0| |decode_ctc0.3_asr_model_valid.acc.ave/test_clean|2620|281530|98.9|0.6|0.5|0.4|1.4|47.6| |decode_ctc0.3_asr_model_valid.acc.ave/test_other|2939|272758|97.6|1.4|1.0|0.7|3.1|63.0| ### TER |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err| |---|---|---|---|---|---|---|---|---| |decode_ctc0.3_asr_model_valid.acc.ave/dev_clean|2703|82834|95.2|3.5|1.3|0.5|5.3|48.2| |decode_ctc0.3_asr_model_valid.acc.ave/dev_other|2864|76205|91.6|6.4|2.0|1.1|9.5|60.4| |decode_ctc0.3_asr_model_valid.acc.ave/test_1h|7439|159974|26.2|48.4|25.4|15.0|88.8|98.0| |decode_ctc0.3_asr_model_valid.acc.ave/test_clean|2620|81195|95.6|3.2|1.2|0.5|4.9|47.6| |decode_ctc0.3_asr_model_valid.acc.ave/test_other|2939|78676|91.6|6.2|2.2|1.0|9.5|63.0| ## ASR config <details><summary>expand</summary> ``` config: conf/tuning/train_discrete_asr_e_branchformer1_1gpu_lr5e-4_warmup5k.yaml print_config: false log_level: INFO drop_last_iter: false dry_run: false iterator_type: sequence valid_iterator_type: null output_dir: exp/asr_train_discrete_asr_e_branchformer1_1gpu_lr5e-4_warmup5k_raw_wavlm_large_21_km2000_bpe_rm3000_bpe_ts6000 ngpu: 1 seed: 0 num_workers: 4 num_att_plot: 1 dist_backend: nccl dist_init_method: env:// dist_world_size: null dist_rank: null local_rank: 0 dist_master_addr: null dist_master_port: null dist_launcher: null multiprocessing_distributed: false unused_parameters: false sharded_ddp: false cudnn_enabled: true cudnn_benchmark: false cudnn_deterministic: true collect_stats: false write_collected_feats: false max_epoch: 100 patience: null val_scheduler_criterion: - valid - loss early_stopping_criterion: - valid - loss - min best_model_criterion: - - valid - acc - max keep_nbest_models: 10 nbest_averaging_interval: 0 grad_clip: 5.0 grad_clip_type: 2.0 grad_noise: false accum_grad: 4 no_forward_run: false resume: true train_dtype: float32 use_amp: true log_interval: 1000 use_matplotlib: true use_tensorboard: true create_graph_in_tensorboard: false use_wandb: false wandb_project: null wandb_id: null wandb_entity: null wandb_name: null wandb_model_log_interval: -1 detect_anomaly: false use_lora: false save_lora_only: true lora_conf: {} pretrain_path: null init_param: [] ignore_init_mismatch: false freeze_param: [] num_iters_per_epoch: null batch_size: 20 valid_batch_size: null batch_bins: 120000000 valid_batch_bins: null train_shape_file: - exp/asr_stats_raw_rm_wavlm_large_21_km2000_bpe3000_bpe6000/train/src_text_shape.bpe - exp/asr_stats_raw_rm_wavlm_large_21_km2000_bpe3000_bpe6000/train/text_shape.bpe valid_shape_file: - exp/asr_stats_raw_rm_wavlm_large_21_km2000_bpe3000_bpe6000/valid/text_shape.bpe - exp/asr_stats_raw_rm_wavlm_large_21_km2000_bpe3000_bpe6000/valid/src_text_shape.bpe batch_type: numel valid_batch_type: null fold_length: - 150 - 150 sort_in_batch: descending shuffle_within_batch: false sort_batch: descending multiple_iterator: false chunk_length: 500 chunk_shift_ratio: 0.5 num_cache_chunks: 1024 chunk_excluded_key_prefixes: [] train_data_path_and_name_and_type: - - dump/raw/train/text.rm.wavlm_large_21_km2000 - src_text - text - - dump/raw/train/text.ts.en - text - text valid_data_path_and_name_and_type: - - dump/raw/dev/text.ts.en - text - text - - dump/raw/dev/text.rm.wavlm_large_21_km2000 - src_text - text allow_variable_data_keys: false max_cache_size: 0.0 max_cache_fd: 32 allow_multi_rates: false valid_max_cache_size: null exclude_weight_decay: false exclude_weight_decay_conf: {} optim: adam optim_conf: lr: 0.0005 weight_decay: 1.0e-06 scheduler: warmuplr scheduler_conf: warmup_steps: 5000 token_list: - <blank> - <unk> - ▁ - S - ▁THE - ▁A - ▁AND - ▁TO - ▁OF - ED - ▁IN - ▁I - T - N - ▁HE - ING - D - ▁WAS - E - ▁THAT - ▁IT - Y - '''' - ▁HIS - M - I - A - LY - ▁HAD - ▁YOU - ▁FOR - ▁AS - ▁WITH - K - ▁BE - ▁HER - R - ER - U - ▁BUT - ▁NOT - RE - ▁IS - ▁SHE - ▁ON - P - ▁AT - L - G - ▁SO - ▁ME - H - C - LE - O - ▁KA - ▁NO - TA - LI - ▁HIM - LA - RI - ▁WE - ▁THEY - ▁ALL - ▁NA - ▁MY - ▁BY - ▁HAVE - ▁O - ▁THIS - AN - ▁WERE - NA - IN - ▁WHICH - ▁DI - NG - RA - ▁MA - ▁AN - ▁FROM - NE - ▁ONE - MA - ▁SAID - W - ▁DE - น - ▁RE - AL - '2' - ▁OR - B - TE - TI - RO - ▁THERE - ▁DO - SE - ▁WHEN - ▁SA - ▁MAN - ▁HA - KA - ▁THEIR - LO - ES - CE - ▁WOULD - NI - ▁C - ▁B - ▁E - ▁WHO - ▁PA - ▁WHAT - ▁UP - ▁IF - SI - LL - EN - ▁THEM - US - ▁ARE - KU - UN - ▁BA - '7' - ▁OUT - ▁เ - 'ON' - ▁BEEN - MI - VE - HA - WA - OR - า - ▁LA - TO - ▁SI - IL - DE - ▁WILL - ▁COULD - CK - ด - UR - AR - GE - DI - KE - ▁M - ▁UN - ST - ▁TA - ▁MO - ▁NI - '1' - GA - SA - MO - ▁MI - ▁YA - ม - ▁MORE - ME - ▁INTO - BA - ▁มี - PA - Á - ▁LIKE - ▁SOME - F - DA - ▁BU - ย - TH - ▁MU - ▁CON - ▁THEN - ▁SE - อ - ง - KO - ก - ▁YOUR - ▁NOW - TED - ▁VERY - ▁CAN - ▁LITTLE - ATION - ▁DA - ▁DID - ▁GO - ▁PI - V - ION - LU - ▁KO - IT - YA - '5' - ▁CO - ▁WA - ▁HAS - ▁ABOUT - ▁TIME - ▁NE - ▁THAN - ▁SEE - ▁KNOW - ▁TU - KI - ▁G - ▁RA - MP - Р - ▁K - ▁SU - RU - ABLE - X - ENT - ▁OVER - ▁LO - ▁TI - PI - ▁ANY - BE - ร - AH - ▁ของ - ▁BO - AT - ▁WELL - CH - Ê - PO - ▁LONG - ▁MEN - ▁UPON - ▁OTHER - ▁GREAT - TU - PE - '3' - ▁PO - ▁TWO - ะ - ▁ONLY - IR - CO - ว - RY - ▁T - ▁OUR - EL - ▁D - ▁AFTER - BO - ▁DOWN - ▁ST - ▁TE - ▁OLD - ▁SHOULD - ▁LI - ▁MADE - TION - ITY - MAN - MENT - ▁BEFORE - А - ▁MISS - ▁GOOD - ▁ครับ - ET - BU - TING - ▁อยู่ - ▁DAY - ANG - و - NESS - CA - ▁WAY - ▁DIS - ▁PE - HE - ▁US - Z - CI - OUS - OL - JA - ▁SUCH - ▁COME - ▁EN - Ი - É - BI - IM - ▁U - ▁HO - '4' - VER - ▁CAME - ERS - ▁HOW - ▁MUCH - ▁ค่ะ - ▁GA - ▁KE - ▁MAY - WE - Ა - ▁DU - UT - ▁HI - เ - ▁PRO - ▁WHERE - Л - ▁JA - Н - ▁NGA - ▁MISTER - ▁BACK - ▁NU - ▁THESE - EST - HO - ▁EM - NT - ▁NEVER - Г - ▁EX - È - Í - О - ▁KU - ▁MUST - ▁JE - ▁ละ - ▁WI - ে - FUL - ▁THINK - Ე - ▁EVEN - ▁BI - ▁JUST - GI - ▁HU - AP - 'NO' - ▁SAY - া - ▁RO - บ - ▁ITS - TY - ▁GI - ▁MAKE - ▁LE - LED - ▁ขาย - ▁HOUSE - ▁OWN - ▁THOUGHT - Д - ه - ล - UNG - QU - ▁FIRST - ▁DON - VI - С - ▁AGAIN - ▁L - DO - ి - ▁OH - ▁P - AM - KAN - IC - ▁บาท - ▁KI - ি - ▁MIGHT - AD - AK - ▁WENT - ATE - ่ - ▁HIMSELF - ISH - ▁AM - ▁THROUGH - Q - ▁VA - Ó - UL - VA - GO - YO - ৰ - À - ▁HAND - ี - LING - LESS - Ò - ▁VI - ▁FA - ▁FAR - ▁F - ▁PER - ANT - QA - NY - ్ - CHA - ู - AS - ▁EVERY - ▁HERE - ▁HEAD - TER - ▁GET - Î - RED - ▁W - ▁N - ▁LIFE - ▁หนึ่ง - Т - YE - ▁TOO - DER - MU - ▁THOSE - ▁CA - ▁WITHOUT - ▁EYES - ▁OFF - ▁MOST - ▁AWAY - ส - IS - Ო - М - Е - ▁MANY - র - GU - OK - ▁NEW - ▁SAW - ANCE - ▁ห - UM - ন - WI - ▁YOUNG - ো - ி - ل - ▁STILL - ▁RU - ▁BEING - ু - ▁UNDER - ▁NIGHT - ر - ▁RIGHT - ম - ▁LAST - У - ▁PART - ▁TAKE - ▁FACE - ▁TELL - HI - ▁PLACE - ▁PUT - ▁FOUND - J - Მ - ▁YET - ▁WHILE - ▁PEOPLE - Х - ▁NOTHING - AY - ▁GRA - OM - ্ - ▁FI - ▁WORK - MB - ▁ต - ▁JU - ▁THREE - FF - ARD - SU - ா - ▁LOVE - ห - ▁THOUGH - ې - ▁PRE - ZE - ▁MAR - ই - CU - '8' - TON - ▁FE - ▁ROOM - ▁PU - HU - TOR - ▁TRA - จ - DU - TIC - ▁SP - ▁SAME - JE - ం - ا - ป - ▁ASKED - VO - ▁LOOK - 的 - PER - ▁EVER - ▁YE - ENCE - ▁HEART - ▁ก - Ş - ต - າ - SON - ▁LEFT - ้ - ▁FATHER - ▁ANOTHER - ▁GOT - ▁LET - ▁ส - ▁TH - ▁ขวด - ▁CAR - ุ - NDA - ZA - ▁SHALL - อง - ▁BAR - ▁V - ▁ONCE - MBA - ท - ▁ALWAYS - JU - RAN - ▁ห่อ - IST - ▁WHY - ی - ิ - ▁SEEMED - ID - Ы - CHE - ▁TOOK - ▁JO - ▁GIVE - ນ - ▁ค - ▁MOMENT - ▁BECAUSE - ค - SO - ▁โ - IA - ▁DOOR - ▁MIND - LAN - ▁HOME - OS - ▁END - Э - ▁TEN - ▁สอง - ▁TOLD - ือ - LES - NDE - ▁SHA - LD - Ვ - ▁H - ONG - ▁LOOKED - IVE - ▁HEARD - EK - ▁SOON - ▁MOTHER - NING - ▁SOMETHING - ้า - ▁APP - ่า - ን - ত - ▁LIGHT - IAN - ▁THINGS - ▁QUE - AU - ▁KING - IES - ▁WANT - ▁อาดอ - ▁CH - Თ - ా - ▁IMP - ▁HIGH - ▁THING - ▁CHA - TEN - ▁SIDE - ▁GOING - ▁NAME - ▁PAR - ZI - Ì - ▁FIND - Რ - ດ - ราะ - ▁WORLD - ன் - ▁PRI - И - RT - FI - ກ - ▁COMP - ISE - І - ▁RI - Ш - ▁그 - م - ▁YU - ี่ - Ọ̀ - ▁EE - ATED - JO - UD - మ - ▁REST - ซ - ు - ▁COM - GAN - ▁CHE - ▁CARE - HAN - ▁WISH - ే - ່ - ት - FER - LAND - ▁WATER - JI - ▁YES - FA - ▁KIND - ▁SHOW - ▁BETTER - แ - ▁FO - ▁LU - ▁WAR - WAN - TCH - س - ▁SIR - ▁KNEW - ▁จ - ▁WOMAN - ▁ร - ▁HARD - BLE - ▁আ - ▁EACH - KIN - ▁ป - AC - ▁ANG - IK - FU - ▁AGAINST - ▁HAVING - ອ - ▁FEW - Ს - ▁GEN - ▁BEGAN - AI - ▁FOUR - ້ - АЙ - ▁SING - ለ - ▁YEARS - ▁AL - త - Й - ▁ENOUGH - ▁SET - ক - LAR - АН - WARD - ▁PRESENT - ▁OPEN - ARY - З - ▁VOICE - STER - IONS - Უ - ▁MIN - ▁WHITE - ROW - ল - ▁EHHE - IE - ▁WHOLE - ▁YO - ▁አ - INE - ব - ▁NOR - ▁BELIEVE - GIN - 你 - ▁GIRL - Ж - OT - ▁J - ▁แ - TRA - ▁SUN - Ẹ́ - ▁HUNDRED - 이 - RON - ▁DONE - TURE - กุ - ฮ - ▁BRA - ▁CALLED - ▁HOPE - ▁AH - ▁MORNING - క - NDI - ▁NEAR - BB - 가 - ▁STE - ▁TAI - ข - ▁STA - ▁Ì - ▁WALK - ▁EL - ▁SEEN - ▁BETWEEN - VED - ▁นิ - ▁บ - ▁FORM - THER - ▁TRI - ▁CLOSE - ም - ANA - ▁STATE - VING - ▁FELT - ▁CHI - ል - YI - WO - IP - Ლ - ▁อ - ▁POWER - TAN - 고 - ம் - WU - ЭЭ - 어 - ДА - ▁มา - ▁GU - IF - ▁HERSELF - FOR - ’ - ▁HALF - বা - ▁TOWARD - Ọ́ - ▁BOTH - ▁হ - ▁POINT - ▁À - ▁AMONG - ▁DOES - ວ - ▁HOWEVER - ▁ALSO - NYA - ▁TURNED - KUNA - ▁POOR - IGN - ▁COURSE - ▁JI - ▁PERHAPS - ▁NG - ▁GE - HAM - ▁কি - ห์ - ▁ORDER - ▁SEA - ▁REPLIED - ▁QUITE - ▁OL - ▁MATTER - ▁MYSELF - TIVE - ு - ▁SURE - 一 - ▁SPEAK - Ө - Ọ - ITE - RING - LT - ▁AR - IOUS - ▁CRE - ํา - ல - KING - CY - TSI - ช - ▁ক - URE - ▁SMALL - ▁GOD - ▁สาม - ▁น - ▁EST - ▁PERSON - ▁GAVE - వ - ▁KEEP - ▁CU - ▁ALMOST - Ú - NCE - TIN - EM - MEN - ▁บู - ▁PEN - COR - Ṣ - OD - ▁MER - IGHT - Ნ - ▁SON - FT - ▁ท - 다 - ▁PLAY - ▁Х - ▁TER - VAL - ت - MER - ▁WHOM - ▁NEED - TUR - ▁WHI - ์ - QUE - ๊ - BY - ▁DEAR - ບ - NGO - ▁DES - ງ - OP - CHI - ▁POR - RIES - 不 - ▁TOGETHER - ▁UNTIL - DAY - Ç - ው - ຫ - ▁HANDS - ▁SINCE - ▁MON - BER - Š - ▁LAND - ي - ▁CHILD - ▁FEET - FULLY - ▁অঁ - ▁NEXT - ▁ANYTHING - ▁WO - 我 - ▁สิบ - ▁BEST - ▁GENERAL - ▁FIVE - ▁WORDS - ▁DAN - 는 - ▁BLACK - АР - ื่น - ▁FACT - ▁SAT - Დ - ▁BOY - డ - ▁LAY - Ხ - ▁MEAN - PP - ▁BROUGHT - LAI - ▁ALONG - AGE - ▁STOOD - ▁WOOD - ய - Ü - RIN - CENT - ▁IYA - 지 - ▁LEAVE - AKE - ▁FRIEND - В - ▁SEN - ▁LARGE - ▁DAYS - ▁SUB - ▁SAN - อินทรีย์ - ▁نه - ▁HORSE - ▁HELP - মা - ▁HEAR - ▁CONSIDER - IYA - ▁RUN - ▁AIR - ▁CHAR - ▁CALL - ▁LIVE - ▁ROUND - ARI - PU - ▁READ - ▁FULL - ▁SIX - ▁USE - DY - ▁DAR - ▁JOHN - ๋ - ▁OU - THI - দ - ▁MONEY - RIC - কে - ▁CHO - ANE - ▁หม - МА - NCH - ▁TWENTY - ே - PUN - UP - ▁FIRE - ▁MASTER - ▁NATURE - นา - RESS - ມ - య - স - TRI - ▁LESS - ট - লে - ▁SENT - ENG - WAY - ▁BEN - ▁LAW - ▁LAN - ▁LOOKING - Ẹ̀ - ▁GUA - ATIONS - ▁CE - ▁RATHER - ▁FEAR - ▁WORD - ▁GLO - Ẹ - ▁SHORT - ▁VO - ▁FAIR - RS - UK - ▁LANG - ▁MIS - LAM - ▁FL - LER - LLA - ▁IDEA - ▁স - АА - พ - ரு - ▁CASE - ▁ᲓᲐ - VES - ▁COUNTRY - ▁INDEED - Გ - ▁PASSED - Ė - PELA - ▁INTEREST - ALLY - ▁QU - ▁PAS - ▁SOUND - เล - ▁ບໍ່ - ስ - TTER - ن - ▁SÍ - NU - い - IZ - ▁FALL - ▁PLAN - ITIES - ん - ▁CRIED - ▁CAP - ▁COUNT - ▁INTER - ▁QUESTION - ่ง - RANG - ▁GROW - ర - ছে - 에 - ▁UNCLE - ▁พ - LLOW - ▁TAKEN - ▁REAL - ▁Б - PING - ▁LADY - ் - ICAL - NED - টা - ▁ไ - ▁க - ▁GONE - ลา - LIN - ▁ACT - ▁THOUSAND - 하 - GRA - ▁OTHERS - MPA - ▁เจ้า - ▁REASON - ▁DOCTOR - LANG - ▁না - ▁AROUND - ▁CLEAR - ▁ব - IANN - SHED - ▁CERTAIN - Ù - ▁SH - ILY - ▁WHOSE - ▁ANSWERED - ▁我 - খ - ▁THEMSELVES - ▁ᲰᲝ - ▁DEATH - ▁RAN - ▁TRUE - ▁ነው - ▁WINDOW - ▁WIFE - Ž - ▁BEHIND - 有 - ▁CHILDREN - UG - ▁ᲠᲐ - ▁BROTHER - ▁NGI - UC - ▁REALLY - ▁ЮМ - ▁TEA - ب - প - ▁PRA - ▁啊 - ▁STEP - TES - নে - ▁GROUND - ิน - ▁TILL - RAY - MENTS - DURING - ZO - CUR - ▁WOMEN - ▁APA - ▁OFTEN - ▁PAN - CHO - ড - FIN - ▁ร้อย - ▁DOUBT - ▁TALK - INA - ▁LETTER - ▁KAY - ▁د - ▁RED - YAN - ▁ล - ▁ARM - ▁SIGN - ▁EAR - AW - ີ - ▁ALREADY - ▁KAN - AUGHT - বে - ▁WONDER - ▁PUR - ▁م - ▁А - HON - ▁ORA - ▁FOOT - ▁BOOK - HAR - ▁FELL - ▁WATCH - তে - ▁HOLD - 是 - ▁STREET - ▁GRE - ▁NÍ - ▁LEG - ▁KON - ▁FLA - ▁สุ - ▁প - ก้า - న - ▁THUS - ▁FINE - PHE - ▁БАЙНА - ▁BECAME - ▁MANNER - LEY - DEN - TERN - ▁SHI - ▁SIGHT - ▁LORD - ▁PARA - ดา - ▁TOWN - SIDE - ▁น้ํา - GUE - ▁BODY - IH - ▁DIDN - ▁FEELING - ▁KWA - ▁WON - ▁VE - ో - ▁ตาก - ▁STRONG - ▁CANNOT - WIN - ▁RETURNED - ▁ЗА - ▁PAIN - ▁PAT - লা - ▁EIGHT - ▁ALONE - ▁BED - స - ถ - ARA - ALI - ▁EVERYTHING - FE - NDO - ▁BIG - ▁แม่ - ▁ILL - க்க - PR - ▁COMING - ▁HAT - Ờ - ▁GIVEN - ▁SECOND - ต๋ - ДЕ - KEN - خ - በ - ÑA - MBI - ▁EZ - Ą - ▁ABOVE - চ - ద - ச - Ğ - ▁REMEMBER - ้อ - TANG - ▁DEAD - ▁OB - ▁你 - ▁MEET - กร - ▁ک - ▁LINE - ▁BEAUTIFUL - Ქ - ▁EXPECT - ▁SLEEP - ▁SEVEN - LAH - PAN - GEN - ▁DARK - ▁CI - ె - Қ - ▁IMA - ▁SUPPOSE - П - ▁EVENING - ▁EYE - UH - PAS - Ც - ▁BER - ▁CITY - ▁FELLOW - ▁HELD - ▁CAUSE - ▁HUMAN - ▁POU - IG - ▁PH - த - DRA - ነ - చ - ያ - ▁MET - ▁ROSE - ▁ART - ▁FEEL - SAN - ▁AC - ▁TURN - ▁FREE - ▁তো - নি - শ - ▁SCHOOL - ▁SOMETIMES - ▁ப - ంట - ▁HOUR - อย - ற - ▁PIN - ▁OO - ▁FORCE - ▁YEAR - ▁CUR - ▁SISTER - ک - ▁UM - ▁UNDERSTAND - ▁DREAM - IYOR - ▁DEEP - ▁SAYS - ▁HAIR - ▁DRAW - ▁STRANGE - জ - ▁LEAST - ▁KEPT - ▁SPOKE - ▁PASS - ይ - ▁হ্যাঁ - ▁OBSERV - จํา - ่ะ - บา - ী - ▁กล้วย - ▁CAPTAIN - ▁Q - Კ - ▁DRESS - ᲕᲘ - ▁SUR - ர - ▁FISH - ▁BAD - ▁FAMILY - PPED - ▁BIR - হ - ᲘᲡ - ▁WALL - ▁BEAR - ASI - นี้ - ▁BECOME - ▁LEARN - ▁ও - ᲛᲐ - ம - ▁OBJECT - ▁ÀWỌ - ▁MM - ้าย - ▁UNA - ር - ▁তা - ▁TABLE - ▁OG - ▁SORT - SHIP - ▁WHETHER - ▁MAKING - ش - ▁PLEASE - ப்ப - ▁MAG - ▁NUMBER - ▁BON - 도 - RUS - CAL - స్ - ▁ES - ړ - 在 - ▁GREEN - Я - ና - ▁DIFFERENT - ▁MOUNTAIN - ▁EARTH - ULA - ▁OFFICE - KHI - ▁ANSWER - ▁WIND - ▁LAUGH - ህ - ▁ซาว - ▁EITHER - ▁FRIENDS - ▁YANG - ங்க - ▁SUDDENLY - ᲐᲠ - ГҮЙ - お - ▁PAY - ▁BRING - ▁WITHIN - ▁RETURN - য় - ▁VISIT - ▁EH - ன - ▁TR - ▁CHURCH - ▁ตรา - ▁BESIDE - ▁BAL - ิง - ▁RING - ▁PRINCE - ▁SPIRIT - ▁ITSELF - ▁THOU - ▁STORY - ▁PAST - ▁NGE - PORT - க்கு - வ - MBO - ▁LOW - গ - 아 - ీ - ▁DAUGHTER - ▁வ - IBLE - ▁SY - LIK - ฟาร - ▁SEVERAL - ์ม - TEL - ▁ELSE - ▁LOST - ▁เกือ - ▁AKO - ▁ROAD - ▁FUN - ລ - ▁SAM - ▁APPEARED - ▁HILL - ▁NÓ - ▁HAPPY - ▁CHU - TIM - ▁POSSIBLE - ▁REC - หม - ▁БОЛ - ▁USED - 라 - ▁SNOW - ▁CUT - ▁RIVER - ▁ASK - ▁ښه - ▁GLAD - د - ▁WEEK - DAN - ▁Ა - IZED - ▁ห้า - ▁UH - ▁ANO - ▁สี่ - ▁STAR - ▁SCR - ໂ - の - ▁PLA - AWA - BAN - ▁COLD - ▁STAND - ັ - ▁SUBJECT - ▁او - ▁WAIT - ▁CONTINUED - ▁FLOW - GON - ຮ - ▁TROUBLE - ▁아 - ని - ▁CHANCE - VIN - ӨӨ - Ñ - ደ - ர் - மா - KEUN - ▁TAN - รี - YU - ঁ - Ả - ▁BLUE - ▁JOY - ▁LISTEN - ▁DESIRE - য়ে - ᲐᲜ - RÍ - ▁LATE - ▁ใน - ▁REACHED - ▁KNOWN - ▁SKI - อบ - TRO - ெ - ▁LÀ - ▁দি - ริ - ▁LEAD - AG - ও - โ - ▁SAVE - ▁AGE - ▁MEANS - ▁ته - WN - ▁QUI - ▁KHÔNG - ▁BUSINESS - ▁FUR - ▁FOLLOWED - LLY - Ч - መ - 서 - ▁COURT - ▁PETER - ▁TREE - ▁SOUL - ▁GRAND - ▁IR - ோ - Ô - ▁EIGHTEEN - ▁THEREFORE - ▁DANGER - ຍ - THOUGH - ▁WILD - LIGHT - ▁NORTH - ▁SAK - あ - ట - ▁MARK - ▁RICH - '0' - ▁เป็น - ▁EXCEPT - ▁GARDEN - ДЫ - ▁WANTED - ▁ACROSS - う - ▁আছে - ▁የ - ▁TOUCH - Ɔ - ▁خو - ▁PERFECT - ULI - ▁NYA - ▁CERTAINLY - WAR - ▁LONGER - ผ - KHU - ▁HUSBAND - ▁OCCASION - ▁BILL - ▁SEEM - ▁ENGLISH - ላ - ▁HẼE - ▁거 - ІН - ▁ARMS - 을 - ▁CHAY - ▁পা - ▁PRINCESS - ▁FRA - IO - ▁CHARACTER - ▁DIFFICULT - ▁OUGHT - ▁SHIP - HIN - ▁ఆఁ - ▁ໄປ - MBER - ไ - ЭН - து - TIK - ЫН - ▁QUIET - ENS - INI - IAL - ▁COL - ز - ብ - ▁ஆஹ் - ▁ถุง - Ტ - ゃ - ▁PRETTY - ▁VIEW - แก่ - ATIVE - KHO - ట్ - LÉ - ▁Л - ৈ - ▁REMARK - ▁SUFFER - ▁ข้าวหอมมะลิ - ยา - ▁TIMES - UX - ▁SECRET - ▁SWEET - ▁OKE - ▁SENSE - ▁READY - ▁DISCOVER - ▁REGARD - ▁CARRIED - য়া - ▁RID - ดี - ▁CHANGE - К - ▁ĐI - ▁ເອີ - ▁چې - ▁COMMON - RAL - ▁ААА - ▁SIN - ▁К - ▁EFFECT - ▁போ - ▁MÀ - ▁INDIAN - เพ - ▁系 - ▁LIVED - ▁LATER - ▁PAPER - ະ - 게 - ▁MHM - ▁TÍ - GUA - ▁จะ - 爱 - ▁SAD - ان - WELL - ▁PROF - ▁BAK - ▁MONTH - ▁CAST - বি - BIL - ▁ప - ▁NATURAL - ▁ADDED - ▁EAT - ▁TRU - ▁NGO - ▁NANG - ▁FRONT - ▁TSIT - ▁HUNG - ▁MANG - ปลา - ▁น้ําผึ้ง - ▁RESPECT - ▁SUCCESS - ▁บ้าน - ▁BROWN - ANGAN - ANI - ▁TAKING - RAC - ABLY - ศ - ▁PUBLIC - ▁SURPRISE - ▁BREATH - か - ▁NH - నా - LUNG - ▁PARTY - KAR - ▁THANK - ▁나 - ຄ - ▁กล่อง - ▁STAY - ▁ที่ - ▁TRANS - ▁IKI - ది - ▁CAMP - Ấ - ▁แดมอ - 기 - 은 - ▁TRUTH - OUGH - ▁SOUTH - ▁TRIED - ▁START - BAR - DHI - ▁PARTICULAR - ▁PURPOSE - ▁SAL - ▁EQUAL - లే - ▁ป้อ - ▁FRO - ▁GAL - PHI - Შ - ▁PREPAR - ▁ANA - ▁WARM - ▁DIRECT - ้ง - ▁ENTERED - ENED - TTLE - ▁ИӘ - ተ - 니 - ▁እንደ - ▁USUAL - ▁STONE - யா - ▁BANK - ▁RECEIVED - ▁FORWARD - ▁AMA - ▁CÓ - ້າ - ▁TREES - ▁GUN - ▁CRY - ▁SUGGEST - ▁แปด - ▁FIGURE - ▁COMFORT - ▁PLAIN - ந்த - ள - RATION - ▁SOFT - ▁THY - ▁ENTER - HOT - ISM - ▁HAYI - ÚN - ▁PAPA - BILITY - EVER - ▁গ - னு - ▁WALA - 就 - ▁BLOOD - ▁POSITION - டி - ▁SAFE - ILE - UB - IAU - ▁GIRLS - ▁এই - พัน - ▁BEYOND - ▁COMMAND - ▁PROMISE - ᲕᲔ - ▁LIVING - ▁MANA - ▁HOURS - ЫП - 리 - బ - ▁ТИЙМ - た - ▁ENGLAND - ▁Ọ - ▁BAI - ▁কর - ▁DELIGHT - గ - ை - ວ່າ - ໃ - ▁PANG - ▁Ừ - Ų - に - WEN - ▁이 - ▁ААН - TUK - NCHIK - ▁AGO - ▁MAIN - ▁BELL - ▁ᲒᲐ - ▁SER - ▁OPENED - ሽ - ▁አዎ - ▁БА - ட - ▁DOING - HANG - ้อง - ▁TREAT - ▁SANE - ᲚᲘ - ▁PALE - ▁پ - 了 - ▁EXPERIENCE - ▁CLASS - ప - FO - ▁বল - PUT - ▁SIT - ▁SERVICE - ▁ENJOY - ▁CHIEF - ▁เคย - ÀN - ▁FER - ▁AGREE - ▁SAYING - ▁ఆ - ▁REMAIN - ▁KANG - డు - ▁FOREST - ▁ข้าว - HING - ▁PLEASURE - な - ▁WORTH - ▁COMPANION - ▁FAST - ▁CARRY - ▁MAL - HOOD - ▁MILES - ก่อ - ▁STRUCK - ▁یې - ▁HELLO - ▁FIGHT - ▁DEAL - ▁SEAT - 都 - ▁BEAUTY - ▁CROSS - ▁SILENCE - ▁INSTANT - க - ▁NDE - ட்ட - তো - HOO - ▁อุ้ย - Û - ▁ข้าวกล้อง - ИН - ตร - 去 - 好 - ▁CROWD - ▁نو - THE - ▁BOYS - ▁BAY - RAH - ▁NATION - ▁SAIL - রে - CHU - ▁BAN - ທ - ▁รับ - し - ర్ - ار - KIM - ▁ยา - IFIED - ▁QUA - ▁ໄດ້ - Į - ▁BRIGHT - ▁НЬ - ▁ZA - ▁ᲐᲠ - ▁MINUTES - ຊ - ▁SAKA - ▁FINGER - ᲓᲐ - ▁IHE - ▁SANG - ▁ACCOUNT - ▁SERVANT - ▁YOURSELF - ▁ຢູ່ - ረ - ▁ກະ - ▁کو - ▁УУ - САН - ▁STANDING - ▁ABLE - ▁ROCK - ▁NEARLY - КЕ - ▁BIT - ะห์ - ீ - ሰ - て - 到 - ▁নাই - ▁THROW - ҚА - ▁কা - 면 - ▁ATTENTION - ▁CONDITION - ▁MOUTH - ▁TRAVEL - ▁را - ▁که - ▁FAT - ▁NON - ตี - ▁SMILE - ▁YOUTH - য - เมียน - ▁PICTURE - ▁FURTHER - ▁BOAT - ▁NAA - ▁VEN - ▁ТА - ட்டு - ▁APPROACH - ▁ຕ - ▁EARLY - ▁HAPPENED - EG - จังหวัด - ሁ - 人 - ▁ปุก - ื - ▁IMMEDIATELY - ▁FLU - ఎ - ▁DRIVE - LOG - ▁GREW - NTEN - ማ - ▁OPINION - ▁COMPANY - ▁PRAY - GGED - ▁YON - ▁BOW - ▁FORTH - ▁EAST - ▁НЭГ - ま - ▁NEITHER - ▁MMHM - ▁ສ - ติ - Ბ - หย - ▁THOR - ▁NINE - ▁ROLL - ▁NONE - ▁ACCEPT - ค่ะ - ▁GOLD - ▁CHAIR - ▁SEEMS - ▁FOLLOW - RIP - ษ - ټ - ▁FLOOR - ▁GLANCE - DDING - ను - KONG - ▁ขอ - ้ว - à - ▁RỒI - ▁THIRTY - ▁THIRD - SCRIBE - ▁WIDE - ▁GATHER - ▁ÇI - ▁THICK - แต - ▁TAIL - গে - ໍ - ▁AFFAIR - 요 - 해 - ▁FRESH - ▁HEAVEN - ได้ - ▁BEAT - না - ▁STOP - ▁MAMA - TAQ - ▁MINE - ల - ▁MUSIC - ▁LENGTH - ▁НЕ - ▁COMPLETE - ▁GRACE - ▁HABIT - CKET - ТЫ - ৰি - ОО - こ - ▁SHOUT - ▁STOPPED - ▁FILLED - ັນ - ▁GUARD - ▁TRO - HOR - ▁QUEEN - ら - ኝ - ▁AFRAID - わ - ▁CLOUD - ▁دی - కు - ▁UTA - ິ - ூ - ▁EVIDENT - き - ▁CREATURE - ▁WOUND - ▁STARTED - ▁HUNT - ▁UTTER - 나 - ته - Ế - ▁DOG - วัน - ▁FIFTY - ▁ไป - ▁SAINT - ZZ - ▁ANNE - ▁FIT - ▁MOON - న్న - ฆ - 个 - ๊ะ - ▁যা - ▁CONTINU - СА - ▁PRESS - ራ - と - く - ▁SELF - ▁PIECE - ▁OKAY - ▁MAH - ▁VER - ▁KORO - ▁HALL - MBE - ▁SN - ▁LIE - ▁STAP - 过 - غ - ▁EXCLAIMED - ▁ЮУ - ▁ATTEMPT - 心 - ▁PROCEED - ▁GUESS - ▁YEN - ګ - ▁GOVERNMENT - ▁REPEAT - తా - ▁BIRD - ▁พื - ▁EXPRESSION - ІҢ - ግ - 唔 - ▁INSTEAD - ▁BREAK - ▁SILENT - ▁APPEARANCE - దు - ▁SPRING - ▁WONDERFUL - ພ - Ạ - ▁EXPLAIN - ▁RESULT - ▁ANIMAL - ▁БИ - LẸ̀ - TSIN - ▁BORN - ▁GRAVE - หา - ▁MASS - ▁ТЭГЭЭД - ▁แสน - 想 - ▁ESCAPE - ▁هو - రా - ▁SITTING - ▁LOVED - ครับ - ▁நா - ▁OUTSIDE - ▁হয় - ຈ - ยัง - ຂ - ฟักข้าว - ▁ขนม - ▁เข่า - ▁MOVED - ▁WEST - ▁GEL - BANG - ▁TRY - ች - ከ - ▁IMPOSSIBLE - り - ▁CORNER - ▁LONDON - ▁DEMAND - ▁WHATEVER - NGGO - লি - 한 - 天 - ▁COVERED - ▁ДЭЭ - CLOCK - ▁TEARS - ▁ERÊ - ▁MAKA - ▁JANE - ▁JOIN - RENG - ంది - ும் - ேன் - ▁เม็ด - ▁DETERMIN - ▁MADAME - ▁PROPERTY - ▁WRITE - ▁HALO - ▁SUIT - PANG - ▁PATH - ▁EXPRESS - ▁BROKEN - TSO - ▁এক - ▁MEASURE - ▁ATTEND - ▁TALKING - ▁XWE - ອງ - లు - ▁POCKET - แก - ᲠᲔ - ТА - ▁BAND - ▁APPEAR - ▁POSSESS - ▁PERSONAL - Ц - ▁هغه - МЕН - ▁WINTER - ▁SCARCE - ▁FÈ - ▁HAPPEN - ▁እን - ገ - ▁ACCORDING - ▁CIRCUMSTANCE - ▁ปิ๋น - ▁FRENCH - ▁CÁI - ▁ATTACK - ▁SHARP - ▁ROMAN - ছিল - BORU - DUL - ▁MWEN - ▁LAUGHED - ▁ЖА - ▁REMAINED - ▁SERVE - え - も - Ń - ▁กระป๋อง - 마 - ▁VILLAGE - ち - ▁AFTERNOON - ▁БАЙГАА - ▁VALLEY - ▁MARRIED - ▁SHORE - ▁POPUL - ▁FORGET - రు - ▁FOOD - ▁THÌ - ▁QUICK - ▁LAID - บัญชี - Ề - ▁EFFORT - ▁HAPPINESS - ▁MAJOR - ▁DISTANCE - ▁FRANK - ▁هم - ▁STORM - ▁PERCEIV - ▁BOUND - ▁PLACED - ▁ARMY - ลัด - ድ - ▁کښې - ▁உம்ம் - ▁ไม่ - ▁ISABEL - ▁WRONG - ▁BLOW - ▁BELOW - ▁BOX - ▁БАР - ▁TAR - ▁RACE - ال - டு - ภ - ▁คุกกี้ - พิเศษ - ▁PROBABLY - 要 - ▁QUARTER - ▁ADMIT - ▁FAITH - ▁GENTLEMAN - ▁SKY - వా - ▁دې - ปา - GGER - কা - ▁YEAH - ▁MARY - ▁TÓ - Ố - ▁PLEASANT - ▁SOCIETY - ▁คัด - హ - さ - ▁GROUP - ▁STRAIGHT - 着 - จาย - การ - ▁FORTUNE - TSAI - ข้าว - ITUDE - หอมมะลิ - ▁STRENGTH - ▁ມັນ - Б - ட் - ▁ENTIRELY - ▁NECESSARY - ▁ҒОЙ - 야 - 있 - ▁DINNER - ▁DREW - ANGA - ▁MEANT - కి - ▁QUICKLY - ᲔᲑᲘ - ▁AMERICAN - ண - ▁SEND - หนองคู - ▁INFLUENCE - ▁BEGINNING - ▁ຊິ - ▁CHAPTER - ▁EASY - ັກ - ▁BROKE - ▁TRAIN - ▁REACH - າຍ - つ - ধ - 사 - ょ - ▁SCENE - ▁PULL - ▁น้อง - ▁GIVING - তি - ▁SLIGHT - ▁COLOR - ▁MEMBER - HOOK - Წ - Ұ - ▁PRODUCE - ▁SILVER - ▁PAUS - ▁DIRECTION - ▁WAITING - กล้อง - ไห้ - ▁AUTHOR - ▁DREAD - ▁HISTORY - ▁SINGLE - ▁BATTLE - ▁SHUT - ГЕ - Ắ - ▁CONVERSATION - ▁ESPECIALLY - ▁ນີ້ - 까 - ొ - ▁EASILY - ▁BREAD - ▁PEACE - ▁OBLIG - ▁FLY - ▁MORAL - ▁ACTION - ฟ - ▁TERRIBLE - ▁தான் - ▁REQUIRE - ▁به - ▁ซอง - లో - ᲐᲡ - నే - ▁ده - ▁АЛ - ▁MILL - ▁AWAKE - ▁STRANGER - ชาย - ▁دا - ▁HARM - ААД - ▁TURNING - ▁TRYING - 들 - ▁HEAVY - 会 - ▁EAGER - ▁አይ - ▁GAME - ▁MBAE - ▁RUSH - వు - ▁LUCK - กลุ่ม - ▁จีพลัส - ▁แห้ง - ▁SIMPLY - ▁SIMPLE - ண் - ▁BELONG - ▁وا - ▁CONTENT - ▁БАЙ - ▁KILLED - ▁REPORT - ▁KUR - ▁SPAR - ▁SICK - ▁LOUD - NGGAL - ▁BAŞ - ▁LAKE - ▁JOURNEY - ▁আৰু - োৱা - ▁ANXIOUS - ▁BILONG - ▁STICK - له - ▁LIPS - ТЕ - IOH - ப - ้ย - ູ - แม่บ้าน - ▁วังภูหมอก - జ - ธ - ▁DECIDED - ▁PALACE - ▁BURN - ▁LAGI - ▁NJE - ▁MAID - ▁MOVE - รินทร์ - ታ - ወ - ▁ADDRESS - ▁STREAM - ▁EVIL - ▁IMAGINE - ▁SLOWLY - ▁CHANGED - னா - ▁REPRESENT - ▁যে - ▁MENTION - ▁ก็อด - ▁FOLLOWING - ▁CATCH - ร้อง - IDAK - ▁MARRY - ▁SUPER - ▁CONCERN - ▁SEARCH - ▁FAVOR - ▁TEMPER - ▁ปลาร - ▁HANDSOME - ల్ - জা - ▁แท - LÚ - ▁THIN - ▁อา - ▁PASSION - ▁SHAPE - ▁ຫັ້ນ - 보 - ▁HÍNA - ▁SUMMER - ▁CIVIL - ▁PRESENCE - ▁SERIOUS - ▁SHOP - ▁SMILED - ▁SPOT - ▁MOTION - KHUAN - ▁AUNT - ▁DUTY - ▁หก - รีบู - Ệ - ▁SUFFICIENT - ▁СОЛ - ▁আমি - ▁SHADOW - ▁BROAD - ▁MISTAKE - గా - เค - ᲨᲘ - ▁ALLOWED - ▁SHOT - ᲓᲘ - ▁GAIN - ▁MINUTE - রা - ▁INDIVIDUAL - ▁ARRIVED - ▁MARRIAGE - ▁COUSIN - ▁SLAVE - ▁ASSIST - ▁อะ - คร - ▁UYA - ▁WEAK - วัด - ▁TWELVE - ▁DEPART - ▁RAISED - TSHU - ▁TRUST - ▁SUDDEN - ▁CHRIST - เบอร์ - ▁UNDERSTOOD - ▁DEGREE - で - ▁HONOUR - ▁GLASS - Ң - CARA - ▁LOWER - Ẽ - ປ - Ფ - ▁CONSCIOUS - ▁เจ็ด - よ - 내 - 안 - 得 - ▁NÁÀ - ▁SUPPORT - ▁NARROW - ▁BATH - ▁KILL - KOH - ▁SPENT - ح - ▁REFLECT - ▁น่าม - క్క - ▁WELEH - ▁FRANCE - ▁CALM - ื่อ - ้ม - ว่า - กัด - ▁INCREASE - ▁FRI - ▁HONOR - ▁FIRM - ▁GOLDEN - ▁POST - ỚI - ▁LATTER - ▁YONG - ▁GRASS - ▁PÉ - BÛ - 으 - ▁เกษตร - ▁ŞEY - লো - ᲑᲐ - ธนาคาร - ▁ADVANTAGE - ▁FASHION - ▁SWORD - 래 - ▁সেই - ▁ENEMY - ▁VARIOUS - ▁NASIDA - ▁SOCIAL - ▁TASTE - ▁ᲡᲐ - ▁BITTER - ▁MOVEMENT - สุวรรณ - ▁เติ้ล - ▁அவ - ▁ຫວາ - 说 - ▁DEVELOP - み - ▁MURDER - ▁LADIES - ▁YORK - ▁ARRANGE - ▁YELLOW - ▁PURSU - HREW - ไชยศิ - Ū - Ჩ - ▁กระปุก - ▁CONDUCT - ▁STRETCH - ▁PREVENT - ▁VICTOR - ▁SITUATION - ▁FINALLY - ▁মই - ▁RELATION - ອນ - ▁ভাল - ีผล - ▁ห้าง - ᲜᲐ - ▁MARCH - ▁TENDER - ЕЙ - ▁MILLION - ున్నా - Ĩ - ▁DECLARED - สมุนไพร - ▁KNOWLEDGE - ▁DROPPED - ▁ມາ - ▁PROPOS - ▁RISE - ▁RULE - ▁กะ - ▁INVIT - Ь - ږ - ณ - ▁ЖАТЫР - উ - บือละห์ - 네 - ▁CARRIAGE - ▁GAYOD - МЫН - ல் - ఏ - ▁ปะ - ای - ▁POND - หุ้นส่วน - 시 - Ơ - Ღ - ▁EVERYBODY - 일 - や - 多 - ▁ລະ - ▁LEAVING - ▁UWIS - ▁دي - దా - SCRIPT - FOLD - ্যা - ూ - ▁근데 - 那 - ▁COLLECT - ▁ANCIENT - ▁PRISONER - ▁RAPID - న్నా - ▁په - ▁DANCE - ของดี - เปอร์เซ็นต์ - ባ - ▁ESTABLISH - ▁என்ன - ▁DISAPPEAR - ▁JUDGE - ▁FAINT - 里 - ኔ - 거 - 来 - 로 - 下 - ফ - 能 - ญ - ج - ف - 를 - Ở - 上 - 오 - 자 - 只 - 没 - 么 - ዚ - ቀ - 为 - 구 - 时 - 这 - 었 - ع - چ - Ồ - る - 情 - 也 - ভ - Õ - ষ - 만 - 인 - ສ - ښ - Ộ - 啲 - 너 - 때 - 무 - 했 - 开 - 又 - ቃ - ክ - が - ொ - ై - 自 - ኛ - Ữ - 哦 - 대 - 여 - は - 边 - پ - Ị - 啦 - 知 - 수 - 远 - 地 - 还 - ひ - 后 - め - 再 - ሱ - Ầ - 같 - 无 - 可 - 려 - 생 - 제 - ຖ - Ớ - 如 - 주 - ሳ - 见 - 话 - 되 - 走 - Ε - っ - 起 - 让 - 데 - Ჰ - ຜ - 像 - 样 - Ę - ጋ - ৱ - ష - 간 - ኮ - ጣ - す - Ể - 看 - 几 - 点 - ጥ - 听 - Ზ - ዳ - ঐ - ຸ - ሉ - 生 - 의 - ق - ቤ - ፈ - 러 - 런 - 回 - ふ - 以 - የ - 정 - れ - ຕ - 道 - 嘛 - 而 - じ - ໄ - Ɛ - Ủ - ろ - ど - Ợ - 出 - Ừ - 感 - 원 - 말 - 세 - ね - 却 - だ - 年 - ዛ - ډ - ሺ - 对 - 日 - 두 - ず - 음 - 笑 - 系 - 소 - 风 - け - ぴ - 分 - 呢 - 든 - 모 - 慢 - 手 - 眼 - 相 - ሚ - ঠ - ণ - 真 - ዋ - 别 - 最 - 애 - ぎ - ቸ - 드 - 랑 - 울 - 차 - ぐ - ^ - ஷ - ሩ - 左 - 할 - ፍ - そ - 头 - ጠ - Ụ - 嗯 - 산 - 운 - ঝ - ካ - 放 - 겠 - ং - ሀ - 谁 - 明 - 间 - 비 - 학 - び - 우 - 카 - 定 - 己 - 늘 - 전 - 더 - ዝ - Ỏ - 많 - 离 - 개 - 星 - Č - Ý - 海 - 상 - Ჯ - 달 - 미 - せ - 然 - ص - ஜ - 之 - 觉 - 很 - 成 - ゆ - ሄ - ሪ - ኩ - 타 - 花 - 부 - ば - 前 - 世 - 和 - 太 - 光 - 把 - 금 - 물 - 스 - 저 - 처 - 次 - 当 - 中 - ຶ - 家 - 未 - 각 - 름 - 막 - 봐 - 신 - 白 - 노 - 已 - ዬ - 언 - ழ - 空 - 住 - 럼 - Ё - 两 - 梦 - 做 - ط - ሌ - 咪 - 度 - 냥 - 던 - 동 - 란 - శ - 温 - 落 - 经 - 给 - Ằ - 月 - Ჭ - ቱ - 流 - 먹 - 望 - 等 - 大 - 小 - 变 - 动 - 讲 - 雨 - 날 - 알 - 약 - 장 - ご - 美 - ຟ - ቅ - 发 - 面 - 길 - 바 - 히 - 失 - 子 - 色 - 걸 - Ổ - 路 - ዐ - む - 同 - 꼬 - 봤 - 치 - 声 - 留 - 每 - 抱 - 带 - 快 - Ǹ - ሥ - Ỉ - 信 - 先 - 老 - 难 - 건 - 디 - 반 - 파 - 方 - 曾 - 泪 - 晚 - አ - 打 - 总 - 十 - ئ - ۍ - ቢ - Ỗ - 굴 - 르 - 응 - 期 - 他 - 所 - 言 - ቶ - 拥 - 歌 - 伤 - 问 - 云 - 更 - ض - 从 - 忘 - 올 - Ự - 少 - 意 - 长 - 怕 - 界 - 身 - 乐 - 永 - 계 - ኑ - 念 - 野 - 살 - ሮ - 于 - 现 - 被 - ぼ - ฤ - ሻ - ቻ - Ặ - 买 - 山 - 님 - 른 - 워 - ৌ - 교 - 直 - ฉ - 些 - 候 - 것 - 냐 - 밤 - 실 - 와 - 좀 - 유 - 喜 - 쿠 - 欢 - 水 - ዜ - 电 - 遇 - 난 - 맞 - 배 - 속 - 않 - 진 - 짝 - 화 - ঙ - ጊ - 呀 - 哪 - 怎 - 위 - 중 - 算 - 微 - 依 - 青 - べ - 清 - 返 - 매 - 별 - 솔 - 줄 - 랄 - 夜 - 才 - 完 - ሬ - 但 - 即 - 忆 - 愿 - 문 - 방 - ሞ - ቆ - 钟 - 轻 - 暖 - 何 - 许 - ሎ - ぽ - 背 - 누 - 因 - 行 - 단 - 돼 - 명 - 엔 - 직 - 집 - 청 - 痛 - 深 - 春 - 实 - 终 - ఓ - ቺ - 본 - 빛 - 새 - 입 - ฝ - 受 - 口 - 터 - ざ - 그 - 安 - 근 - ໋ - 停 - 怀 - 车 - 쳐 - 트 - ሊ - ሙ - ሲ - Ẫ - 节 - 갑 - 갔 - 년 - 눈 - 린 - 분 - 柔 - 千 - 向 - ஸ - 单 - 事 - ቼ - ኳ - 喺 - 待 - 食 - 강 - 레 - 예 - 절 - 죠 - 容 - 错 - 느 - 种 - 反 - 静 - 唱 - 火 - 近 - ژ - 苦 - 회 - 루 - 버 - 불 - 왔 - 甜 - 飞 - 满 - Ũ - ጀ - 此 - ໊ - ጂ - ፊ - 够 - 热 - 께 - 록 - 몇 - 적 - 져 - 쫌 - 행 - 刻 - 牵 - 音 - 往 - 双 - 法 - ణ - ሠ - ኪ - ጉ - ጭ - 用 - 结 - 며 - 영 - 외 - 조 - ':' - ̣ - ሜ - ቂ - ぬ - 겨 - 귀 - ధ - ぞ - 儿 - 哭 - 越 - ̀ - 跟 - 假 - 叫 - 阳 - ጎ - 作 - 原 - 思 - 竟 - 답 - 偷 - 它 - 气 - 沉 - 理 - 细 - 转 - 重 - 높 - 밖 - 선 - 식 - 씩 - 연 - 잎 - 종 - 피 - 久 - 入 - 外 - 脸 - 靠 - 城 - 醒 - 找 - 早 - 写 - 偶 - 友 - 该 - 渐 - 곱 - Ф - ຝ - ፋ - Ử - 吧 - 告 - 긴 - 재 - 프 - 녀 - 성 - 테 - 三 - 装 - 夏 - ほ - 角 - 寻 - 睡 - ஐ - ኋ - 과 - 求 - 玩 - 香 - 곰 - 머 - 빨 - 쪽 - 코 - 콩 - 亲 - 今 - 女 - 紧 - 온 - 호 - 默 - 机 - 勇 - 请 - 首 - 关 - 掉 - 全 - 岁 - 活 - 감 - 공 - 六 - 蓝 - ዴ - ጅ - ጆ - ጤ - Ỡ - 使 - 包 - 啱 - 嚟 - 她 - 画 - 钱 - 雪 - 국 - 글 - 쁜 - 업 - 息 - 随 - 与 - 底 - 烟 - 滴 - ぜ - 步 - Ю - 比 - భ - げ - 学 - 将 - 希 - 正 - 闲 - ሷ - ቡ - ዕ - Ễ - 味 - 尽 - 整 - 条 - 解 - 进 - 슬 - 용 - 체 - 케 - ጃ - 场 - 웃 - 似 - 红 - 计 - 疯 - 语 - 엎 - 万 - 必 - 敢 - 旧 - 秋 - 街 - 切 - 排 - 遥 - 담 - ኸ - ዱ - 力 - 秒 - 결 - 된 - ฬ - ቁ - ኖ - 五 - 埋 - 平 - 懂 - 旁 - 漫 - 飘 - 렸 - 받 - 쉬 - 슨 - 양 - 철 - 침 - ጫ - 습 - 片 - 绝 - 四 - 害 - 守 - 约 - 书 - 傻 - 北 - 否 - 酒 - 季 - 残 - 照 - آ - 累 - 꾸 - 발 - 천 - ሸ - ኞ - ዙ - ፌ - 乌 - 吗 - 始 - 市 - 应 - 恨 - 独 - 线 - 诺 - 透 - 격 - 경 - 따 - 맛 - 몸 - 썰 - 였 - 질 - 크 - 후 - 工 - 迷 - 토 - 丽 - 影 - 句 - 恋 - 需 - 黑 - 散 - 奔 - 啊 - 们 - 张 - 目 - 亮 - 忍 - 群 - 鱼 - 强 - 挂 - 넘 - ث - ሂ - ሴ - 倾 - 管 - 荡 - 갈 - 십 - 엉 - 커 - ৎ - 另 - 晴 - 穿 - 若 - 谎 - 闹 - 阵 - 목 - 월 - 편 - ৃ - 公 - 处 - 干 - 合 - 坐 - 怪 - 易 - 站 - 认 - 狂 - 至 - 体 - 提 - 笔 - 收 - 阴 - 追 - 高 - ぱ - 二 - 断 - 球 - 耳 - 诗 - 遍 - 配 - 预 - 복 - 짜 - ظ - ఖ - ኒ - 习 - 冷 - 特 - 졸 - ሔ - Ჟ - 任 - 休 - 便 - 哩 - 字 - 报 - 改 - 灵 - 烧 - 神 - 纸 - 联 - 部 - 롱 - 색 - 씨 - 추 - 悲 - 浪 - 肯 - 西 - 东 - 初 - 半 - 局 - 脑 - 距 - 缘 - 聊 - 非 - 承 - ዎ - 灯 - 彩 - 惜 - 接 - 交 - 保 - 孤 - 运 - 代 - 圈 - 憾 - 差 - 纯 - 连 - 逃 - 九 - 其 - 南 - 号 - 江 - 演 - 톡 - 혼 - ఐ - ዶ - ጓ - ぶ - 乱 - 决 - 叶 - 响 - 奇 - 尾 - 屋 - 林 - 模 - 训 - 论 - 迹 - 靓 - 났 - 등 - 떤 - 앞 - 통 - 희 - 传 - 八 - 化 - 曲 - 窗 - 表 - 证 - 립 - 송 - 태 - 台 - 恩 - 楼 - 并 - আ - 熟 - 怨 - 送 - 景 - థ - 떡 - 右 - 坏 - 娘 - 本 - 足 - 通 - 隔 - 利 - 名 - 常 - 数 - 碎 - 门 - ذ - ቹ - ጡ - ፕ - Ỹ - 仔 - 倦 - 剩 - 封 - 尘 - 执 - 晨 - 泡 - 猫 - 痕 - 谅 - 谓 - 超 - 跳 - 轮 - 醉 - 망 - 붙 - 순 - 옥 - 옹 - 움 - 증 - 쪼 - 축 - 팔 - 럽 - 七 - 莫 - 选 - 항 - 噢 - 妈 - 尔 - 灰 - 躲 - 刚 - 握 - 零 - 挡 - 死 - 贴 - 杯 - ఆ - 围 - 绕 - 拿 - 丝 - 悠 - 旅 - 百 - 止 - 观 - 吻 - 喵 - 堂 - 怜 - 懒 - 戏 - 草 - 顺 - ௌ - ዮ - 골 - 딩 - ቋ - 呼 - 存 - 摆 - 斗 - 油 - 般 - 视 - 점 - 향 - ঃ - ౌ - ጪ - Ẩ - 企 - 养 - 哥 - 妙 - 惯 - 搞 - 擦 - 木 - 朵 - 波 - 注 - 淡 - 班 - 英 - 茶 - 贵 - 迎 - 锁 - 题 - 饭 - 马 - 骨 - 관 - 깔 - 끔 - 둥 - 떻 - 랐 - 룩 - 먼 - 민 - 벽 - 셨 - 얀 - 억 - 임 - ฎ - ፒ - 达 - 闪 - 颗 - 긋 - 嘴 - 撑 - 男 - 短 - 突 - 续 - 荒 - 识 - 诉 - 黄 - 低 - 折 - 舍 - 寄 - 朝 - 祝 - 课 - 挥 - 瓶 - 礼 - 幻 - 战 - 试 - 琴 - 닷 - 伞 - 剑 - 卷 - 吸 - 哈 - 惊 - 拒 - 梁 - 燃 - 租 - 第 - 羽 - 脚 - ጌ - 品 - 喝 - 漂 - 铁 - 메 - 밥 - 키 - 페 - ̩ - ሶ - ቄ - ዪ - Ẻ - 享 - 价 - 伯 - 傍 - 冬 - 升 - 吞 - 国 - 急 - 房 - 抬 - 指 - 新 - 昏 - 替 - 服 - 涌 - 游 - 滚 - 田 - 眸 - 码 - 篇 - 芳 - 豆 - 退 - 避 - 酸 - 鲜 - 궁 - 깐 - 댁 - 덕 - 뜨 - 벗 - 베 - 석 - 숲 - 역 - 짓 - 쭉 - 쯤 - 찜 - 출 - 클 - 폰 - 활 - ቴ - 급 - 댕 - 력 - 준 - 합 - ቲ - 争 - 余 - 吵 - 唯 - 尝 - 旋 - 甘 - 놓 - 충 - 乎 - 盛 - 纷 - 辉 - 偏 - 挽 - 洋 - 立 - 颠 - 忙 - 藏 - 暗 - 跌 - 倒 - 含 - 层 - 古 - 格 - 临 - 极 - 脏 - 酷 - 魂 - 资 - 吃 - 根 - 毛 - 沙 - 碰 - 舒 - 蝶 - 辜 - 院 - 修 - 染 - 柠 - 烽 - 移 - 血 - 途 - 颜 - 魔 - 릴 - 법 - 패 - ሏ - ሯ - ኗ - ዊ - ዓ - ፏ - Ỳ - 伏 - 借 - 共 - 冒 - 冲 - 功 - 叹 - 君 - 圆 - 垂 - 寒 - 寸 - 座 - 扬 - 抗 - 拉 - 换 - 揾 - 教 - 斑 - 浮 - 添 - 港 - 潮 - 烈 - 牌 - 牙 - 瘦 - 眶 - 砰 - 祷 - 穷 - 答 - 纪 - 绿 - 翻 - 肉 - 胜 - 苍 - 象 - 赖 - 辰 - 逐 - 镜 - 限 - 须 - 餐 - 骑 - 骚 - 鸦 - 겁 - 넌 - 놈 - 닐 - 될 - 뜩 - 렴 - 론 - 롭 - 쁘 - 심 - 씬 - 악 - 짐 - 쩔 - 탈 - 탕 - 튼 - 판 - 현 - 셋 - 쟁 - 환 - 唐 - 性 - 涯 - 物 - 珍 - 疼 - 缠 - 夕 - 설 - 쳤 - ፎ - 卖 - 套 - 汤 - 良 - 솜 - 瑞 - 稳 - 缺 - 伴 - 唤 - 序 - 归 - 挑 - 翅 - 薄 - 咸 - 义 - 件 - 列 - 勾 - 嘿 - 属 - 岔 - 广 - 弹 - 掩 - 搭 - 欠 - 猜 - 符 - 腕 - 阔 - 낙 - 펼 - Ï - 华 - 嫁 - 幽 - 抓 - 暂 - 烂 - 珊 - 疲 - 翼 - 触 - 逆 - 闻 - 킨 - 商 - 흥 - ً - ቭ - ቮ - ዩ - ፅ - Ḿ - Ỷ - 丘 - 严 - 介 - 伪 - 位 - 冻 - 净 - 凉 - 刘 - 刺 - 博 - 厚 - 呵 - 嘈 - 团 - 壳 - 奏 - 姐 - 婆 - 宝 - 宫 - 宴 - 密 - 尊 - 川 - 店 - 延 - 引 - 徒 - 悦 - 惑 - 抖 - 抵 - 抹 - 拆 - 拖 - 拼 - 救 - 暴 - 束 - 校 - 款 - 毫 - 洗 - 测 - 湖 - 湿 - 灌 - 煌 - 熄 - 熬 - 犹 - 环 - 皮 - 盖 - 眠 - 票 - 秘 - 稻 - 窝 - 纵 - 绍 - 缝 - 考 - 者 - 舟 - 虹 - 警 - 讨 - 词 - 负 - 躯 - 载 - 逝 - 逼 - 量 - 针 - 际 - 陷 - 馆 - 鬼 - 麻 - 黎 - 龙 - 걍 - 껏 - 꿀 - 끊 - 낀 - 낼 - 똥 - 램 - 럴 - 렀 - 맨 - 몽 - 변 - 브 - 블 - 뿡 - 샘 - 싹 - 써 - 접 - 졌 - 줬 - 즈 - 짹 - 쨌 - 쫙 - 찡 - 채 - 컵 - 켜 - 틀 - 티 - 팅 - 폼 - 품 - 픈 - 读 - 견 - 멘 - 뻐 - 헤 - ዞ - ፆ - シ - 主 - 刹 - 智 - 朗 - 权 - 炼 - 盏 - 릭 - 승 - 份 - 加 - 孩 - 摇 - 欲 - 造 - 金 - 隐 - 菊 - 黏 - ́ - 伟 - 婚 - 弄 - 招 - 毁 - 毕 - 激 - 踢 - 鼻 - 嗅 - 妞 - 尖 - 异 - 弦 - 弯 - 彻 - 烛 - 甸 - 眷 - 练 - 荣 - 蝉 - 雁 - 骗 - 齿 - 육 - ఘ - 塞 - 帘 - 悄 - 拾 - 搁 - 晶 - 漠 - 竹 - 篱 - 羞 - 肠 - 闯 - 띄 - 뭇 - ቨ - 핀 - '6' - ሟ - ሹ - ዷ - ጮ - ” - 业 - 乞 - 乡 - 井 - 亦 - 仰 - 俯 - 兮 - 兴 - 军 - 凝 - 凭 - 刮 - 剧 - 午 - 卡 - 卫 - 取 - 叛 - 叠 - 司 - 吼 - 嗌 - 困 - 块 - 坠 - 堆 - 堪 - 墨 - 奢 - 妖 - 姜 - 姿 - 嫌 - 嫣 - 宜 - 宠 - 客 - 寥 - 尺 - 岛 - 岸 - 巾 - 师 - 弟 - 弥 - 悬 - 悭 - 悸 - 惘 - 扁 - 扇 - 抢 - 抽 - 拳 - 探 - 推 - 插 - 支 - 文 - 料 - 斯 - 昼 - 暮 - 李 - 枯 - 某 - 栏 - 案 - 桥 - 欺 - 歉 - 沦 - 沸 - 泊 - 泥 - 淀 - 渊 - 源 - 溢 - 滞 - 滩 - 澈 - 澜 - 灿 - 炎 - 烦 - 煎 - 煲 - 狗 - 珠 - 瓷 - 盈 - 盒 - 盘 - 瞰 - 石 - 破 - 碌 - 碟 - 禁 - 程 - 箱 - 糊 - 糖 - 纹 - 织 - 绒 - 维 - 罢 - 罪 - 职 - 股 - 脊 - 腰 - 腾 - 膛 - 舞 - 船 - 茧 - 莞 - 莲 - 菜 - 蔚 - 蛊 - 蜡 - 融 - 衡 - 衣 - 衫 - 袋 - 讯 - 详 - 谈 - 谷 - 购 - 费 - 赔 - 赠 - 趁 - 趣 - 蹦 - 轰 - 辑 - 输 - 辛 - 辣 - 迟 - 逻 - 铺 - 锦 - 闭 - 闷 - 阻 - 附 - 陆 - 降 - 鞋 - 韵 - 顶 - 顾 - 顿 - 颤 - 馈 - 驶 - 验 - 骼 - 鸡 - 鸣 - 鸥 - 麽 - 검 - 궐 - 긍 - 껄 - 껴 - 꼈 - 꼴 - 꽤 - 끗 - 능 - 덜 - 덟 - 둔 - 딜 - 땐 - 떳 - 뚜 - 뚱 - 뜰 - 뜻 - 띤 - 랫 - 례 - 료 - 링 - 맥 - 맺 - 밀 - 범 - 볶 - 섹 - 앨 - 엾 - 옛 - 존 - 줍 - 찌 - 첫 - 춰 - 칠 - 켰 - 쾌 - 큼 - 텁 - 톨 - 플 - 허 - 험 - 헴 - 홉 - 힌 - 봄 - 뻤 - 쩌 - 巴 - 忽 - 愧 - 投 - 柳 - 滥 - 犯 - 调 - 끌 - 值 - 嫩 - 宿 - 废 - 建 - 恶 - 旦 - 板 - 治 - 爸 - 玉 - 疗 - 眯 - 瞒 - 设 - 蹈 - 辆 - 段 - 覆 - 乖 - 梯 - 举 - 힙 - 힐 - 효 - 혹 - 헬 - 퓨 - 탐 - 큰 - 츠 - 책 - 짠 - 잤 - 웨 - 엘 - 엑 - 앉 - 씀 - 썼 - 쌀 - 싱 - 숙 - 삼 - 뿌 - 뽑 - 뻔 - 벙 - 백 - 멋 - 락 - 똑 - 딴 - 뒤 - 녕 - 납 - 김 - 값 - 齐 - 麦 - 鸽 - 韶 - 隶 - 陶 - 阑 - 释 - 逸 - 辘 - 轳 - 赐 - 豫 - 谢 - 诶 - 衬 - 蝴 - 虽 - 虑 - 莎 - 腻 - 肥 - 聚 - 聆 - 翱 - 缤 - 稀 - 积 - 社 - 矜 - 盼 - 痣 - 疆 - 畅 - 甩 - 猪 - 父 - 烫 - 灭 - 溃 - 渴 - 淹 - 淌 - 涩 - 汉 - 樱 - 森 - 棉 - 朽 - 曦 - 晰 - 敞 - 摹 - 摸 - 掠 - 捞 - 挤 - 抉 - 慌 - 愁 - 恒 - 式 - 廿 - 廓 - 宾 - 室 - 宋 - 孔 - 契 - 夸 - 士 - 垫 - 土 - 嘉 - 喇 - 喀 - 啥 - 哼 - 厌 - 勒 - 冚 - 兰 - 兜 - 兑 - 俗 - 伸 - 丰 - “ - Ẵ - ፖ - ፐ - ጽ - ጵ - ጄ - ቪ - ሑ - ຣ - ฑ - 흘 - 핸 - 필 - 풀 - 퍼 - 탁 - 컴 - 춤 - 착 - 찢 - 죽 - 좁 - 읽 - 빼 - 봅 - 병 - 맘 - 땅 - 딸 - 둠 - 눌 - 녔 - 냇 - 낄 - 깊 - 龟 - 骄 - 饮 - 除 - 银 - 逢 - 踩 - 谱 - 衍 - 蜚 - 葬 - 获 - 苹 - 苞 - 芬 - 祈 - 番 - 狼 - 狈 - 渺 - 泣 - 树 - 敷 - 故 - 拜 - 扣 - 憨 - 惦 - 屌 - 备 - 境 - 坡 - 圳 - 嗮 - 喽 - 喧 - 善 - 啸 - 周 - 呃 - 医 - 准 - 充 - 傲 - 倔 - 佛 - 且 - ጩ - ጁ - ሦ - ఛ - 흔 - 햇 - 평 - 팀 - 특 - 캐 - 춥 - 최 - 줘 - 죄 - 왕 - 숨 - 뛰 - 걱 - 额 - 页 - 踏 - 赏 - 贪 - 脆 - 耀 - 翔 - 网 - 继 - 童 - 瑰 - 玫 - 犀 - 炊 - 洒 - 汹 - 欣 - 梗 - 晕 - 晏 - 掌 - 担 - 护 - 徨 - 彷 - 弱 - 奶 - 堡 - 坦 - 兆 - ஈ - 흐 - 휴 - 휘 - 훔 - 확 - 형 - 함 - 포 - 취 - 잔 - 웠 - 옷 - 뼉 - 밑 - 맑 - 득 - 둘 - 늦 - 넓 - 굳 - 갖 - 陌 - 遮 - 逗 - 较 - 赶 - 诚 - 胸 - 绪 - 络 - 精 - 昨 - 慰 - 悉 - 崩 - 奈 - 埃 - 嘞 - 努 - 京 - ฒ - Щ - 폴 - 염 - 빗 - 넣 - 군 - 겼 - 谂 - 膀 - 肩 - 瞬 - 牛 - 桃 - 既 - 帮 - 姑 - 复 - 咋 - 及 - 仿 - 付 - ፀ - 흰 - 투 - 족 - 잊 - 싫 - 뿐 - 밝 - 밌 - 멀 - 릎 - 떨 - 듣 - 됐 - 닭 - 곳 - 袅 - 耐 - 择 - 彼 - 坚 - ኤ - ቷ - 찍 - 완 - 볼 - 벌 - 딱 - 닌 - 낌 - 꽃 - 꺼 - 誓 - 蜜 - 茫 - 持 - 办 - 乜 - ኢ - ฏ - ء - 즘 - 옆 - 뭔 - 끝 - 跑 - 河 - 楚 - 攞 - 或 - 忧 - 弃 - 寞 - 啡 - 咖 - ぺ - ぷ - 황 - 탉 - 참 - 암 - 쓰 - 确 - 寂 - Ъ - İ - 찾 - 잠 - 싸 - 떠 - 당 - 놀 - 끼 - 괜 - 광 - 陪 - 简 - 究 - 찮 - 즐 - 돈 - 睛 - 消 - 匆 - Ẳ - ኘ - ஊ - 푸 - 탄 - 섯 - 빙 - 吹 - ቦ - 았 - 술 - 깨 - 蚊 - 唉 - 哎 - 仲 - 친 - 창 - 잡 - 왜 - 꿈 - 遗 - 福 - 朋 - 힘 - 찬 - 슴 - 몰 - 뚝 - 남 - 阿 - 幸 - 哇 - へ - ጨ - ሐ - 초 - 열 - 랬 - 작 - 畀 - ኬ - 由 - 命 - ኦ - ฐ - ぅ - 싶 - 돌 - 睇 - 啫 - ঢ - 또 - 喔 - 손 - 걔 - 얘 - 빠 - ఈ - ঞ - 얼 - 못 - 喂 - 엄 - ূ - 잘 - 嘟 - 什 - ኧ - 좋 - 吓 - 번 - 람 - ヴ - 记 - 없 - ஞ - ዘ - 잖 - ఒ - ぉ - 咩 - ぁ - 嘅 - 㗎 - ዲ - ஓ - 咯 - ஏ - ஒ - ஃ - ぃ - Პ - 뭐 - 冇 - 렇 - ሆ - ঘ - Ძ - ゅ - ぇ - Ă - ̃ - Ậ - ఫ - 佢 - 咁 - 果 - ຽ - ఊ - ఉ - Ứ - Ყ - ځ - உ - ຼ - Ö -  - ຢ - څ - ఇ - থ - ஹ - Ә - ళ - எ - ఁ - ங - ື - Ư - ஆ - இ - Ғ - অ - ແ - அ - అ - ந - ົ - এ - ใ - Đ - ় - Ү - 丢 - 劣 - 匹 - 哑 - 嗓 - 嗨 - 嘲 - 填 - 宏 - 巷 - 志 - 扔 - 拙 - 桂 - 梨 - 渲 - 潦 - 爬 - 痹 - 签 - 素 - 翘 - 胚 - 腼 - 茹 - 虎 - 蚝 - 衷 - 褪 - 跃 - 逛 - 釉 - 钢 - 锐 - 队 - 饰 - 빔 - 偿 - 凑 - 剔 - 呦 - 增 - 宣 - 席 - 户 - 批 - 披 - 拂 - 拌 - 捧 - 搅 - 昔 - 晒 - 曝 - 松 - 栀 - 桐 - 檀 - 汗 - 液 - 炬 - 瑚 - 稍 - 篆 - 绽 - 聪 - 莹 - 蒙 - 袱 - 贝 - 违 - 뷔 - 呜 - 瑟 - 딪 - 릿 - 멈 - 셔 - 킬 - ఔ - Œ - ڼ - ঔ - ๆ - ሕ - ሼ - ኙ - ኜ - ኡ - ኽ - ዉ - ዌ - ዥ - ጢ - ጦ - ጧ - ጬ - ፂ - ፉ - ፓ - ク - 串 - 丹 - 产 - 亭 - 仍 - 仕 - 仙 - 优 - 估 - 佬 - 侈 - 侍 - 侵 - 俊 - 倏 - 倚 - 催 - 允 - 兄 - 冰 - 况 - 减 - 凡 - 则 - 判 - 制 - 刷 - 剪 - 割 - 助 - 劳 - 勉 - 匙 - 区 - 卅 - 卑 - 卓 - 占 - 印 - 厂 - 历 - 厕 - 厢 - 叮 - 史 - 吊 - 吭 - 呐 - 呓 - 呕 - 咆 - 咛 - 哀 - 哮 - 唇 - 唏 - 啩 - 喻 - 嘶 - 器 - 噩 - 嚷 - 囊 - 园 - 图 - 培 - 堕 - 塘 - 墅 - 墓 - 墙 - 壁 - 央 - 奚 - 奥 - 妨 - 妹 - 妻 - 娜 - 媚 - 孑 - 孓 - 孙 - 宁 - 官 - 宛 - 宽 - 寐 - 寓 - 察 - 尻 - 屉 - 屎 - 展 - 峰 - 州 - 巧 - 帶 - 帽 - 床 - 庞 - 弘 - 形 - 彰 - 征 - 徊 - 律 - 徘 - 循 - 忐 - 忑 - 忠 - 态 - 怅 - 怡 - 恐 - 恙 - 恢 - 恼 - 悔 - 患 - 悴 - 惟 - 惠 - 惧 - 惨 - 惩 - 惫 - 惹 - 愈 - 愣 - 愫 - 慨 - 憔 - 戒 - 扎 - 托 - 扯 - 抛 - 拋 - 拘 - 拢 - 拣 - 挣 - 捱 - 掀 - 掂 - 掏 - 揽 - 揿 - 摔 - 摞 - 摧 - 撂 - 撩 - 敏 - 敲 - 斟 - 旌 - 族 - 旗 - 旺 - 映 - 昧 - 晃 - 晓 - 晖 - 普 - 暄 - 暧 - 曳 - 曹 - 曼 - 末 - 杀 - 杂 - 杆 - 材 - 杰 - 枪 - 柄 - 柜 - 栖 - 框 - 桦 - 桨 - 梢 - 梭 - 棒 - 棠 - 椅 - 槽 - 檬 - 欧 - 母 - 毯 - 民 - 汁 - 池 - 汪 - 汽 - 沁 - 沫 - 沱 - 沾 - 沿 - 泰 - 洁 - 浅 - 浆 - 浑 - 浓 - 浦 - 淘 - 淮 - 渣 - 湛 - 湾 - 溉 - 滂 - 滋 - 滑 - 漏 - 灼 - 炭 - 烁 - 烊 - 煤 - 煮 - 爆 - 版 - 率 - 王 - 玛 - 玲 - 琵 - 琶 - 瑕 - 瓜 - 瓢 - 瓣 - 疏 - 疚 - 疤 - 痒 - 痴 - 皱 - 盗 - 盲 - 眉 - 眺 - 睄 - 矮 - 硬 - 碍 - 碑 - 碗 - 碧 - 祸 - 秀 - 私 - 秃 - 窄 - 窑 - 竭 - 笆 - 筐 - 筑 - 簸 - 米 - 类 - 粉 - 粘 - 粤 - 粥 - 絮 - 繁 - 纠 - 纱 - 绑 - 绘 - 绢 - 绣 - 绮 - 绳 - 绵 - 绻 - 缄 - 缆 - 缓 - 编 - 缚 - 缱 - 缸 - 罗 - 罚 - 羊 - 羔 - 翠 - 耕 - 耘 - 聋 - 肤 - 胆 - 胎 - 胭 - 脂 - 腆 - 腐 - 膝 - 舱 - 舷 - 艺 - 芒 - 芙 - 芜 - 药 - 萍 - 萎 - 蒸 - 蕖 - 藕 - 蘸 - 虚 - 虾 - 蚕 - 蜃 - 蟹 - 补 - 衰 - 袭 - 裙 - 褴 - 褶 - 规 - 订 - 议 - 讽 - 访 - 谊 - 谋 - 谜 - 谣 - 谬 - 豪 - 贞 - 账 - 货 - 贸 - 赤 - 赵 - 跤 - 踞 - 踟 - 蹰 - 躺 - 软 - 辈 - 辩 - 辽 - 迁 - 适 - 逅 - 逍 - 递 - 邂 - 邪 - 邮 - 酱 - 钓 - 铃 - 铲 - 锋 - 镌 - 镯 - 闸 - 闺 - 阱 - 陈 - 隙 - 雀 - 雅 - 集 - 雷 - 霓 - 霸 - 靛 - 鞘 - 颂 - 馨 - 驳 - 骂 - 魄 - 魅 - 鲍 - 鲤 - 鸯 - 鸳 - 鸿 - 鹅 - 鹤 - 鹰 - 걘 - 걷 - 겸 - 곡 - 곤 - 곽 - 굽 - 권 - 극 - 깜 - 꼭 - 꽁 - 꽂 - 꾼 - 꿇 - 꿔 - 뀌 - 낮 - 냈 - 널 - 녁 - 놨 - 뇨 - 눠 - 뉴 - 늪 - 닥 - 덩 - 뎅 - 독 - 돋 - 돔 - 듯 - 딘 - 땜 - 떴 - 똠 - 뚫 - 랗 - 랩 - 량 - 련 - 롤 - 룰 - 룸 - 림 - 몬 - 믄 - 믿 - 박 - 봉 - 북 - 붉 - 븐 - 빅 - 빚 - 빡 - 빴 - 뺄 - 섬 - 솟 - 쇄 - 쉴 - 쉽 - 슷 - 쎄 - 쏘 - 씻 - 앍 - 앎 - 압 - 앙 - 얇 - 얹 - 엠 - 엥 - 옮 - 옵 - 옾 - 욕 - 웅 - 웬 - 율 - 윽 - 익 - 잃 - 잇 - 젤 - 줌 - 즌 - 징 - 짱 - 째 - 쨈 - 쩍 - 쩐 - 쪘 - 쫓 - 찔 - 챔 - 첨 - 총 - 춘 - 췌 - 측 - 층 - 칫 - 켓 - 콧 - 큐 - 킹 - 탑 - 턱 - 턴 - 털 - 텍 - 텐 - 톱 - 퇴 - 퉁 - 튀 - 틍 - 팩 - 팬 - 팽 - 펴 - 픽 - 햐 - 헐 - 혀 - 혔 - 혜 - 혤 - 홍 - 훨 - እ - ጸ - ጹ - 蓦 - 霞 - 넷 - 녹 - 쌓 - 욱 - 택 - 텔 - 표 - 典 - 冠 - 凤 - 啤 - 委 - 庆 - 弗 - 悍 - 惭 - 慕 - 搬 - 斜 - 梳 - 略 - 疑 - 矗 - 航 - 芍 - 芽 - 褛 - 辗 - 迫 - 醺 - 键 - 露 - 鷁 - 专 - 仅 - 克 - 免 - 叙 - 咳 - 嗽 - 塌 - 富 - 峭 - 峻 - 恻 - 拍 - 枝 - 橙 - 涟 - 漪 - 睁 - 砸 - 组 - 羁 - 萄 - 营 - 葡 - 败 - 赴 - 雕 - 颓 - 驻 - 各 - 氧 - ছ - ํ - ເ - ็ - ึ - ั - 伙 - 坎 - <sos/eos> src_token_list: - <blank> - <unk> - 么 - 喤 - 佨 - 叡 - 卐 - 伍 - 乻 - 勀 - 习 - 众 - 亿 - 勐 - 呵 - 偦 - 乖 - 乸 - 伿 - 丆 - 並 - 卭 - 侲 - 亶 - 再 - 丽 - 偯 - 乍 - 乔 - 伴 - 儑 - 倓 - 呔 - 傛 - 厚 - 喉 - 傓 - 別 - 仩 - 仮 - 乜 - 佸 - 今 - 勾 - 勝 - 喭 - 喵 - 入 - 呃 - 俥 - 丨 - 剏 - 喽 - 儻 - 亣 - 劺 - 佃 - 侖 - 傎 - 儈 - 兌 - 做 - 刢 - 俩 - 喩 - 五 - 傶 - 乩 - 傒 - 僝 - 厊 - 几 - 匎 - 俉 - 吠 - 厪 - 侽 - 丩 - 划 - 侧 - 仛 - 呇 - 乣 - 刡 - 仟 - 其 - 兹 - 咎 - 啧 - 从 - 冉 - 俴 - 伾 - 冱 - 倌 - 勠 - 勲 - 叜 - 伢 - 删 - 伻 - 唼 - 儼 - 唴 - 上 - 兏 - 児 - 儝 - 喲 - 丁 - 侕 - 傉 - 且 - 兄 - 卫 - 人 - 伂 - 仏 - 唦 - 匶 - 侸 - 冎 - 吾 - 伎 - 凪 - 北 - 仆 - 劸 - 喍 - 仳 - 凛 - 傲 - 养 - 厹 - 傷 - 仸 - 吉 - 下 - 併 - 勢 - 劆 - 叵 - 儺 - 价 - 吡 - 剧 - 兾 - 侫 - 喃 - 兙 - 俫 - 匨 - 侓 - 佳 - 剚 - 劘 - 倆 - 听 - 丅 - 哲 - 侷 - 同 - 僓 - 剣 - 券 - 匡 - 专 - 侴 - 勥 - 仰 - 咮 - 唬 - 唶 - 吺 - 偂 - 仙 - 喨 - 刹 - 乾 - 主 - 伅 - 兒 - 保 - 叽 - 唊 - 乥 - 哏 - 儕 - 佗 - 刅 - 偪 - 俋 - 價 - 之剂伇 - 侒 - 侙 - 侑 - 卥 - 啡 - 凢 - 傇 - 佣 - 丵 - 偿 - 偒 - 唣 - 匴 - 俯 - 叺 - 哵 - 丢 - 佤 - 俅 - 丹 - 傯 - 匂 - 刊 - 傩 - 匠 - 升 - 叱 - 亇 - 准 - 仯 - 伒 - 句 - 唗 - 亽 - 匛 - 來 - 倩 - 傏 - 傱 - 公 - 哙 - 吹 - 儷 - 喆 - 喎 - 付 - 営 - 勩 - 卣 - 侚 - 刯 - 伝 - 呬 - 侟 - 丂 - 七 - 俪 - 唫 - 刄 - 厃 - 伲 - 享 - 勁 - 吰 - 咅 - 凓 - 倵 - 匮 - 啇 - 口 - 但 - 儾 - 乑 - 厐 - 勱 - 偢 - 呌 - 刔僬于 - 伳 - 冹 - 剈 - 凰 - 丄 - 劫 - 哨 - 乃 - 僧 - 僋 - 嗎 - 厦 - 丈 - 喇 - 亴 - 匊 - 啗 - 儎 - 傕 - 咡 - 亞 - 仈 - 假 - 偀 - 剘 - 古 - 伞 - 伤 - 両 - 京 - 咂 - 儉 - 劑 - 乏 - 冧 - 劬 - 哣 - 刘 - 傮 - 勡 - 只 - 傴 - 卅 - 周 - 倝 - 叹 - 令佋 - 听乬 - 儁 - 仃 - 仓 - 促 - 乛 - 傪 - 丶 - 倢 - 喬 - 亗 - 唢 - 傋例佈 - 傾 - 哅但 - 匩 - 兩 - 凂喐 - 剬 - 勚 - 仲 - 偤 - 儭 - 侻 - 刷 - 啯 - 侶 - 勊 - 含 - 位 - 刭 - 丝 - 停 - 勛 - 危 - 凍 - 倘 - 偳 - 剱 - 俍 - 哈 - 亗勽 - 冊 - 克 - 仇 - 九 - 匹 - 勎 - 冴 - 亡 - 侞 - 万 - 乂 - 佩 - 丌 - 喑 - 備 - ▁丕些 - 冂 - 冯 - 劳 - 倷 - 他 - 丯 - 僊 - 儩 - 勽 - 亅 - 乼 - 僗 - 伊 - 俛 - 呧 - 僯 - 佂 - 傀 - 件 - 厓 - 亥 - 却佻 - 侠 - 僾 - 伟 - 俹 - 凸 - 厅 - 兕 - 厸 - 久 - 僄乌 - 仌养 - 乘 - 匱 - 伄 - 佢 - 啵 - 亯俒 - 倬 - 僔 - 劇 - 啴 - 儲 - 丑 - 吇 - 俦 - 佐 - 僮 - 僙 - 傳偗 - 凶亦 - 傧冨 - 亠 - 僁 - 偭 - 伃 - 哭 - 倮佉 - 味 - 乞 - 啳 - 剅 - 倍 - 伅傊 - 代 - 乀 - 哾 - 临 - 丘 - 佘 - 冟 - 兙勢 - 乤 - 劜 - 凫 - 咗 - 仍 - 傧 - 丳 - 丫 - 侢乲 - 号 - 双仼 - 倲 - 哅 - 仢 - 喗 - 佟 - 乧 - 啨 - 举 - 丼 - 光 - 俜 - 冨 - 兮 - 华 - 伦 - 唾 - 偌 - 丟 - 哊 - 佭 - 俖 - 偮 - 傟 - 喂 - 匢 - 况 - 佹 - 傮剂伇 - 俘亷 - 俰 - 仡 - 乮 - 劰 - 傊 - 刁 - 凑 - 僨俢 - 侰 - 厧 - 俄与 - 产 - 係 - 剰 - 仹 - 乭 - 吞 - 卺 - 剃 - 亢 - 伔 - 僣 - 啅 - 列 - 僤 - 兖 - 单 - 丒 - 即丶 - 叩 - 匔 - 俄 - 倮 - 侺 - 劏 - 励 - 咡倲 - 乙 - 凁 - 匈 - 使 - 啌 - 厀 - 乆 - 僕 - 啃 - 匁 - 仺 - 俵 - ▁咧倒 - 兺 - 亵刀 - 业 - 会 - 伀 - 亷 - 咛 - 傺 - 倞份 - 全 - 匾 - 商 - 呣 - 厄 - 唬厐 - 伐 - 剁 - 凴 - 佝 - 借 - 僰 - 凗 - 份 - 乊 - 剠 - 卛 - 佬刟 - 們 - 俸 - 劽 - 厥 - 吢 - 两 - 侥 - 凿喿 - 仠 - 员 - 咤 - 厴 - 仌 - 叭 - 傜 - 喦 - 凲 - 吙 - 僑 - 偓 - 卫丼 - 卢 - 凎 - 唍 - 供 - 冊俪 - 乿 - 伮 - 唼儻 - 亐 - 僟 - 伬 - 丰 - 啓 - 刏 - 俻 - 偣 - 卤 - 僀僽 - 偔 - 企 - 伜 - 冩 - 倡 - 唸 - 勿 - 乌 - 剛 - 傳 - 亦 - 凣 - 丷 - 冀 - 吧 - 倇 - 凹 - 刎 - 又 - 傰 - 吆亨 - 兢 - 劈 - 剫 - 侐 - 儔 - 呿 - 啝 - 丿 - 凮 - 前 - 卾 - 俟 - 凉 - 伈 - 喊 - 叐 - 乘偭 - 仅 - 作 - 务 - 倎喍 - 咦 - 吞凗 - 喛 - 嗇 - 俌 - 哘 - 召 - 向 - 吚 - 叅 - 儀 - 仴 - 予 - 喉丩 - 卶 - 唹 - 倷乺 - 伯 - 傑 - 厜 - 否 - 倠争 - 喸 - 仉 - 啢 - 佈 - 俆 - 冈 - 咺乱 - 啰 - 亙 - 厐兺 - 倿 - 伶 - 匏 - 叞 - 俗 - 呖 - 剠乂 - 剖 - 后 - 唟 - 哋却佻 - 厵 - 兊 - 右叿 - 呖八 - 之剂啴 - 亜 - 咢 - 初 - 伤傎 - 南 - 剎勖 - 喓 - 八 - 函 - 凼 - 啚 - 侸劤 - 勈 - 匼 - 傽 - 嗃 - 唎 - 儇 - 三 - 偋 - 唽 - 吚匨 - 匕 - 亻 - 倠 - 倿参 - 唏 - 倰 - 和 - 員 - 偶 - 刽 - 呀兆 - 傍 - 傣 - 佖 - 価 - 仒 - 先 - 勡伃 - 丙 - 刌 - 停乛 - 傐 - 佻 - 俅啵 - 倾 - 务儡 - 匦 - 佯 - 傆 - 佑 - 侣喙佚 - 剆 - 匃 - 兼 - 了 - 僄乌丿 - 劤 - 匪 - 卟 - 吭 - 勰 - 勶 - 侨 - 凔三 - 兩俉 - 劯 - 东 - 俓 - 僴 - 凪凢 - 啸 - 刚 - 写 - 务儡吶 - 仨 - 倅 - 倎 - 乢 - 俧 - 偼 - 傥 - 伋 - 咂哙 - 剡亣 - 勿亹 - ▁ - 傅 - 仼 - 唎償 - 与 - 凒乎 - 到 - 乽 - 充 - 凡劍债 - 刂 - 侯 - 哢 - 刨 - 加 - 凶 - 優 - 吆 - 叿 - 剩 - 吡咎 - 合 - 侕侥 - 唖 - 卸 - 厥佘 - 僀付 - 佲 - 唎唿 - 僘 - 卺凟 - 吚冭 - 仪 - 凔 - 卉 - 卓 - 剃唳 - 仑 - 刜 - 兀 - 卯 - 唁 - 倞 - 丞 - 亭 - 劼 - 俀唀卜 - 仯唴 - 亓 - 乱 - 厲 - 叧 - 偍 - 厛 - 偒傰 - 唬兺 - 佥 - 哜 - 傋 - 互 - 予佔 - 丸 - 唘 - 亏 - 侜判你 - 儯 - 仜 - 哥 - 叮 - 吽佩 - 义佣 - 吼 - 儈乾 - 什 - 乖喨 - 传 - 吿 - ▁丕侄 - 偾區 - 伕 - 兓冹 - 偠 - 些 - 依 - 侳 - 喧 - 体亳 - 喒仝 - 台 - 侞伢 - 乫亻 - 偎 - 則倹 - 不 - 厔 - 吋 - 卞 - 呏侯 - 乓丶 - 仳勝 - 亸 - 咓 - 倻 - 倍哥 - 丠 - 偂儣 - 傭 - 乣乩 - 乐 - 働劯 - 喒 - 匞 - 俏咽僤 - 唵 - 劕 - 儗 - 仾 - 兡 - 亮 - 兝 - 勥乑 - 刦 - 俼 - 使侻 - 咿 - 儥 - 丳厵 - 儺乬 - 乷 - 偏 - 僵 - 厭 - 兗 - 丣 - 呀丰 - 侰丌 - 俘 - 偂伃 - 剷 - 叝 - 倜 - 冃咠 - 唰 - 佔 - 儷傊 - 偘 - 侜判 - 亖 - 偩亚呈 - 倖 - 厢 - 傚 - 刳兎 - 仜仏 - 吲 - 勨 - 什南 - 再勾 - 勺 - 乊傕 - 刔僬 - 啫 - 以 - 勵 - 唛 - 叔 - 唜 - 兑 - 唅哢 - 仚 - 住 - 傖 - 佃侽 - 匩倵 - 劮 - 侮 - 唀卜 - 卯呶 - 刋 - 儰 - 呁 - 佾 - 剩况 - 乶冯 - 举乢 - 佃仛 - 哓 - 佬倏 - 叵仅 - 名 - 儤 - 偅 - 仿叿 - 匢兵呅 - 卉哓 - 喊匞 - 匄 - 侹冗 - 哋 - 呫 - 丞啌 - 午 - 叏 - 几乍 - 哋佲 - 啊 - 勛亀 - 僣佴 - 佊叢 - 喷 - 吞凗喧 - 厒 - 僭 - 俑 - 呉 - 侘冚 - 亵 - 侘 - 侜你 - 乚 - 俽 - 乓傍 - 哀 - 云 - 冶 - 刟 - 則 - 作乢 - 侵勊 - 咪佒 - 凾 - 唸万 - 唩 - 仭 - 凟 - 千 - 厒叵 - 勤 - 佦 - 兲 - 哯 - 唿 - 乡 - 侴丼 - 呪傫 - 凼佲 - 吅 - 兪 - 吏 - 卸僗 - 匜儳 - 乒冣 - 和刭 - 哯仿叿 - 侞剁 - 嗃仦刍 - 乌丿 - 倳 - 偄 - ▁丕侄冰 - 勖 - 厬 - 吐 - 凪剏 - 啲 - 史 - 勃 - 僽 - 偫 - 伺 - 刻 - 叞冓 - 刳 - 儏冤啱 - 冗 - 双 - 唩哥 - 俘丏 - 叕俯 - 嗇区作 - 倚优 - 儚 - ▁咧倒丧 - 勬 - 兓 - 乓 - 喱 - 偙丿 - 倐 - 凪呃 - ▁儱 - 啹 - 刐 - 乩乣 - 偄僟 - 偟 - 偈 - 俏咽 - 咊俎 - 咡哾 - 刀 - 喋 - 令呻 - 儚丠 - 为侎 - 侦 - 丗 - 他喊 - 刾 - 厌 - 哸 - 乲 - 咞 - 呁傑 - 冩僋 - 叴伷 - 凑仓 - 僯右叿 - 丧 - 喷吐 - 交 - 冶乞 - 厉 - 亰产 - 唜侮 - 兮佅 - 仱 - 乽丟 - 唩唏 - 伉 - 俜佞 - 叽伻 - 唳 - 伉値匒咕 - 佶 - 剎 - 剄 - 乄 - 僄丿 - 咾 - 佇 - 唅 - 冰 - 亹 - 凼啯 - 伺勐 - 咪仇 - 不佉 - 乂储 - 佣儃 - 凝咽 - 匔丛 - 佴佯 - 卂劮 - 佴 - 吚啾 - 启 - 偬 - 另 - 交吢 - 井 - 劣 - 丯业 - 冁 - 問 - 傋例 - 叜偒 - 优勓 - 儱 - 唇 - 厃吰 - 丐傒 - 唢丠 - 仫 - 凝咽僤 - 啺 - 厱 - 俌叝喳 - 丂假 - 勾呃 - 侹 - 啍 - 儧 - 唬厐兺 - 冭 - 剤 - 兗况 - 刂冂 - 剂 - 介僆唳 - 亘 - 出喳 - 匫 - 俁 - 呈 - 兪勩 - 仑仮 - 吶 - 俐喗 - 咬 - 俻劑 - 俇 - 僓傰 - 吨 - 凝兓冹 - 僎 - 右 - 叭右叿 - 呫咓 - 叔亥 - 仈卟 - 乩吉 - 吽 - 勃仝 - 储 - 喉剏 - 乩伿 - 備侮 - 乺 - 丂停 - 俽刌佐 - 兝亜 - 偍伇 - 兟 - 僎仡 - 代哏 - 介亲劄匏 - 哬 - 喱呧 - 呭厲 - 刑侞 - 僁傪呧 - 佛僽付 - 匘厢 - 丐儃 - 侸双 - 乬 - 冃关唆 - 出 - 倏 - 吋呓 - 僻 - 則倹乺 - 乶 - 兎 - 喰匸 - 僣佴佯 - 凵 - 唇傅 - 卒 - 仝 - 剛伯 - 呭 - 卻 - 哯叿 - 卂 - 劰刎 - 俹俢 - 匙 - 厢件 - 啲傚 - 剽 - 匍 - 偗 - 偙 - 乛俜佞 - 俽刌 - 佐佅厮 - 僔喃 - 丣厑勧仑 - 厔刁 - 佛僽 - 僌 - 共 - 偲 - 促佹 - 厘 - 傞到 - 侶加 - 傮偍伇 - 亰 - 侒几 - 咆 - 佚 - 呃么 - 哒 - 俟举 - 喷侐 - 也 - 厹乑 - 否佥 - 你 - 剒 - 兲儧 - 乪僔 - 伬兎 - 呪丛 - 侸仼 - 吅仂 - 唩啯 - 億勗匍 - 俢 - 偷 - 咺 - 伽 - 历 - 儏 - 厶 - 吟儿 - 下伬 - 仆兠僻 - 乙凛 - 传喰 - 喗傎 - 儚乞 - 冒 - 亵刀凑 - 唥 - 傟兠僻 - 唛偞倽劕 - 即 - 丂儲 - 义刻儃 - 喰 - 倄剟 - 叞佱 - 儃 - 喳 - 么刢 - 僷 - 呹 - 吏俿 - 仁 - 俅兓冹 - 偡 - 个哀偷 - 億勗 - 乫 - 偡卋 - 咔 - 冑 - 傊亞 - ▁咧倒匁 - 冸 - 呐僩 - 冀前 - 俿 - 丗兊 - 剟 - 匔呪佡 - 僄 - 叢 - 匳冊 - 唃兜 - 伊佸伊 - 乼僛 - 丐 - 剞 - 叛 - 兮厮 - 侵 - 予佔凑 - 啗冧 - 兓咽 - 乭卻 - 侺厱呓 - 侇僈哽俪 - 冠哛 - 佉 - 兌剚 - 哽 - 僆唳 - 哞 - 凸儣 - 呭凲 - 債劊 - 呶 - 儊 - 丁佂傪叱 - 勎刨亱傚 - 呭亀 - 俏 - 喰先 - 不僛叐 - 佟劳 - 咬佊叢 - 凧 - 凰冔 - 呐 - 亩 - 个 - 佸伊 - 唥初 - 侊 - 唏儺 - 伉剑倫 - 佛 - 倀 - 侠公 - 丑冭 - 効 - 哯仿 - 呐哾 - 乛俜伄 - 佋 - 啱 - 剰侯 - 傤 - 乯 - 伿侒 - 乙匾 - 勥叽 - 僙咠 - 僽付 - 俴兝亜 - 优 - 咀 - 吉乩 - 偓凎 - 冤啱 - 倚勓 - 劦偁 - 兵呅 - 呏 - 咶劫 - 叒 - 劰刎厲 - 冄 - 伋冺吣 - 儸 - 喱呧叱 - 刄冂 - 冚 - 劄 - 匭 - 俗侵勊 - 剤凊偈 - 凁啨 - 伋亘咆俧 - 凙 - 厥卶 - 仹偙丿 - 唎亹 - 叼 - 呃喤 - 僬于 - 亍伡佒 - 告 - 剏凪 - 厍伧 - 丹兹丹 - 周勚 - 僓俉 - 勰儧 - 似 - 几亴 - 剠厔 - 務側冫 - 促唴 - 偽 - 伊佸 - 伽仇 - 仹偙 - 刿 - 侢 - 伲佟 - 卟剬 - 喝 - 刎凲 - 冓 - 吁 - 咕 - 伦亀 - 傶侽 - 勿償 - 剤凊 - 呪佡 - 啟 - 俦俆 - 哩 - 創 - 仱伟 - 仆勬名 - 刐偓 - 儊吶 - 唵卂劮 - 凸伃 - 傇佅厮 - 仚匄 - 俄件 - 乵 - 傶價 - 判 - 哿 - 儏剞亍 - 倧 - 呭兡 - 凥喓 - 億 - 哨倵 - 伤剘 - 佱 - 判你 - 剁叜 - 人來 - 呥偬 - 偲兙 - 傰伶 - 傯交 - 倌兪 - 偫冗 - 厃仳 - 儏剞 - 俜佞乖 - 仃专 - 么儲 - 傞亙 - 侷仩 - 厩 - 勲來 - 凘 - 呪 - 勽含 - 令 - 侕什 - 一 - 個 - 之剂 - 偋份 - 喡 - 免仑 - 前双 - 吻 - 中 - 厎 - 丙劺 - 侔倷 - 佟乶 - 俀厬 - 偪侽 - 倌业 - 卛冶乞 - 启佲 - 凶仼 - 咞另 - 合卢 - 兕享 - 丯厄 - 剙 - 争 - 凔吧 - 停乑 - 僾喷吐 - 伉剑倫咕 - 勽业 - 佤僩 - 仍唍 - 唫創 - 助 - 倽劕 - 侢千 - 佊 - 凂 - 兾仴 - 倄 - 侪 - 傴侵 - 儂 - 偘啹 - 傹 - 卌 - 么僤 - 兺呃 - 僱呉們 - 唌兩 - 勮 - 博 - 匳 - 冡哛 - 伐厃 - 剘啌 - 兛 - 俴兝 - 咬占啍 - 僈 - 啳卢 - 倊啇 - 乃儤 - 冔 - 価僊 - 伻剏 - 个偷 - 唌 - 呻 - 僨 - 儈儲 - 儓 - 勋 - 冖 - 儿 - 呃勾 - 哋却 - 俰亜僴 - 何匬兙 - 呏又 - 兓啵 - 义刻 - 俽傇 - 傮剂伇啴 - 呀争 - 価儺 - 和勊 - 喛厩 - 侦佖 - 偪习 - 咅傏 - 卜 - 冖否佥 - 劼儿 - 傴侵勊 - 匕佇 - 倽 - 仍僰 - 冻 - 仧 - 亊 - 傟乨 - 偟兲儧 - 俁健傹 - 働 - 儖傤 - 勓 - 剂伇 - 剠侺厱呓 - 凝俏咽僤 - 咬佊 - 伤仁 - 勿亹償 - 僝哨 - 刋儔 - 区 - 僀 - 卺亨 - 勠呀兆 - 剠乂储 - 劓 - 偹 - 却 - 侢勶 - 前劤 - 叠 - 凿 - 乛唬伄 - 偶匪 - 僝匩 - 係侴 - 剖喛 - 劘丯 - 刿倎喍 - 哅但吉 - 唁呣 - 偟兲 - 以丄 - 刑 - 兹亴兹 - 台乕偗 - 兝久 - 僆 - 傝 - 书啈卵 - 則倷乺 - 傓偯 - 么倠争 - 免仑仮 - 唄 - 佔叔 - 儁侫 - 丽勬名 - 凙剫 - 僩 - 伨 - 临唜 - 呥 - 厴厸 - 啫価 - 呿啹 - 唵卂 - 冭匨 - 匜佀儳 - 僛 - 伎倰 - 俋倻 - 喥 - 匸 - 乴偾區 - 哋喒仝 - 唑 - 咿励 - 哛 - 剻 - 侥劈 - 叫 - 唰丄俰 - 剆匨 - 債 - 任 - 呦匥 - 嗉下 - 佡 - 丮 - 伉剑倫丑冭 - 吖 - 偖呴侏偗 - 吿啹 - 兮佅厮 - 侎 - 伨仁 - 俲 - 務側冫哾 - 俶 - 仹偙兀 - 台傳偗 - 僑亨 - 僓偒 - 之倀 - 倷吏 - 儰偹乄 - 佗吡 - 啖喍 - 剦 - 准伊 - 厍 - 兹亴 - 其俐喗 - 劬侢 - 伬呁傑 - 仺卤 - 唴兒 - 儜仪 - 傲员 - 一區 - 伨倎 - 僰劳 - 啫偭 - 兠 - 凊 - 偬伈 - 估 - 俗侵 - 佔凑 - 乤厅 - 凡 - 匋 - 區 - 军 - 剭丮劣丗兊 - 儠 - 凒 - 亂冦卬 - 呉們 - 侉 - 儚倌兪 - 仹匾兀 - 劜咿 - 吘吷伥 - 唖伴唖 - 侵刭 - 伶儺 - 並儲 - 侤 - 休 - 亳 - 叽勥 - 侖勲 - 傇佅 - 佀 - 咁 - 倜倡 - 佫 - 古儉 - 兪佑 - 吹儕 - 來喇 - 吉俅 - 勻 - 呪佡喽 - 喙 - 冱丨冱 - 吏俿凹 - 准咅 - 仆兠乨 - 免 - 偛 - 佸伊佸 - 债 - 俞 - 伈凣厜 - 亸凘 - 厐亢 - 倚 - 专体亳 - 偶匹 - 啩 - 劭亳 - 僧卥 - 內 - 交励 - ▁仗僒 - 呵冃唆 - 偾區倨 - 兖吟儿倞 - 吘买别冀 - 勥叽勥 - 侣 - 倪 - 吿呿兢 - 啇侘 - 仵 - 丁佂傪呧 - 九伴九 - 丘唣 - 串 - 伶僊 - 咞另凫 - 剭 - 佱员 - 冡 - 乿呶 - 亃 - 咘俯 - 乱乧 - 偅亖 - 偫会否佥 - 厑仞 - 唢乞 - 內咐 - 呐僩勊 - 匋勮 - 啉 - 乶伐 - 儕厃 - 凧偷 - 唍以 - 倅代倅 - 咋 - 喿 - 倩佼 - 劬乲 - 値匒咕 - 刻儃 - 優唥 - 佈亠 - 厴卢 - 亨 - 仨刋 - 勨喿 - 咑 - 勼仜 - 吽佩专 - 列叩 - 亠啢 - 僇 - 呡 - 侞吏 - 劙制 - 俫吾 - 偣叛 - 儳 - 兾仳 - 啚丌 - 佸伊佸伊 - 吘买别係 - 唘倡 - 亴兹 - 佘吿呿 - 儼啫 - 兽 - 儮 - 丏 - 呐偏 - 倻俋倻 - 乕偗 - 偟儧 - ▁仗些 - 侇僈哽 - 偞 - 后凣 - 俗哾 - 厘劊 - 伬刳兎 - 儑佑 - 傶习 - 俽仝 - 亭后 - 俯們 - 倜俸 - 刓 - 冑会否佥 - 剑倫 - 喌 - 佐佅 - 偎匨 - 俚下 - 呥偬傅 - 価仃 - 唴伲 - 価伺 - 勥叽伻 - 厤 - ▁咧哒 - 傅仧兟 - 卓僈哽俪 - 呍 - 仱倄剟 - 佁 - 凥 - 亵京 - 伞僁 - 后凣厜 - 善 - 唷 - 僃 - 儡吶 - 喜 - 僱們 - 刨亱傚 - 勽俌叝喳 - 吊傸 - 吿啹兢 - 乑伻 - 匱刂冂 - 侢勈 - 吆卺 - 储吋 - 佅 - 勤刐 - 呇児 - 关 - 仿 - 劰亀 - 哕 - 喎从 - 佞 - 傐唥 - 侊丣 - 厉乗吥 - 咠 - 侓佳 - 哚 - 匦啚 - 俿凹 - 價傶 - 亀 - 侅 - 仅儩 - 乗吥 - 丒勫 - 劇勖 - 儔丙 - 伫 - 凥勫 - 剽哱 - 侲創 - 喐 - 亮刚 - 倞凗 - 么儈 - 哉 - 卓僈哽 - ▁仗侄 - 卂倢 - 俟傉 - 也啚 - 匔傫 - 冄儼 - 伮勤刐 - 厒仅 - 亮刚啺 - 亍 - 买别 - 乾儈乾 - 倐冀前 - ▁仗勯剄冒亊 - 吠凢 - 參冄 - 傟兠乨 - 勼 - 亵京吨 - 剝 - 亼佛 - 倚优勓 - 叴伙劉 - 勺俸 - 乪 - 仍侐 - 匙丗 - 咂匠 - 僁傪叱 - 厄儀 - 哬剈 - 叫冀 - 亞仏 - 儯佋 - 吣 - 刽仱倄剟 - 叛偶匪 - 匂喎 - 儖任 - 厰 - 儖傤任 - 兆 - 僦 - 乍几乍 - 伓 - 偂咬佊叢 - 唒 - 偾 - 临傖 - 勺呓 - 偭仏 - 俾儿 - 匜佀兩 - 伹 - 傺加 - 喨侰 - 侰厍伧 - 匹俫 - 低啱 - 俐 - 劖 - 佒 - 哘伮勤 - 亿僨俢 - 咭仪 - 嗃功伆功 - 乾么乾 - 冪 - 唣删 - 佼 - 仝傇 - 丯含 - 唆 - 咶 - 刑侞伢 - 俘丏偩亚呈 - 几亴几亴 - 二 - 伪 - 剰剈 - 勗 - 侕儇 - 卒佁 - 佂傪叱 - 僯右 - 俓合 - 僙冃唆 - 傍勐 - 厸乞 - 代倅 - 偞倽 - 哜哬 - 傴刭 - 啧儣 - 倿凊 - 亄 - 倯凯 - 倮僛叐 - 俌乯喳 - 乕 - 仱伟剟 - 厄倩 - 偏傚 - 匜佀 - 傋七 - 主剈 - 厸丠 - 伡 - 剭佺劣丗 - 俓哅 - 喱傪呧 - 僳 - 倁 - 傀亼佛僽 - 哩伨倎 - 勇 - 俉啝 - 儣 - 勐乬 - 凮啃 - 僾喷吖 - 俾 - 侺储 - 凴兩俉 - 劦偁劄匏 - 仟儎 - 亽唣 - 亁啶咖 - 傗 - 倴 - 僄但 - 厚乑 - 伏 - 亝俯 - 喇儎 - 冽侗 - 兕刐 - 亭后凣厜 - 咞倬 - 侩 - 准啵 - 傣厸 - ▁丕冰 - 吘 - 偤句偤句 - 喊匮 - 唬伄 - 伈咢 - 刷佱 - 召偹乄 - 侃 - 丁佂呧 - 人伲 - 偣叅 - 伞佂傪呧 - 侕侥劈 - 侇 - 咞另傚 - 刚啺 - 售 - 剚他 - 勫 - 乮啗 - 俽刌佐佅厮 - 凔三吧 - 內喝 - 乫亻凸 - 唰丄 - 厞 - 侕侰 - 偌前 - 唸乭卻 - 俓哅但 - 傎俗刭 - 伬刳 - 剷丳厵 - 剴 - 体 - 丽兠僻 - 丝喦 - 偖 - 亴兹亴 - 临唜侮 - 俤 - 世匙 - 劘喤丯 - 写你 - 伺僊 - 剬俢 - 乚唳 - 俚 - 优冴 - 予吰 - 保伴保 - 何仕 - 唖伴 - 伐厅 - 伉値匒 - 儅丫 - 冤 - 匬 - 伺儺 - 上伲 - 倊 - 係侳 - 为像 - 偢亽 - 咙 - 乔凷 - 刌佐 - 債劊冗 - 唫刘 - 仍吐 - 亱傚 - 剞亍 - 傘 - 倠么倠争 - 单凸 - 丈冱丈 - 倹 - 低呸 - 侂 - 冃 - 呓 - 冝 - 互五 - 几亴几 - 啭 - 劋 - 厴喸 - 匼倽劕 - 冹乾 - 乼不 - 亼僀 - 准冹 - 侑僋 - 値 - 儃侨 - 咘 - 亴几亴 - 丽兠乨 - 仚咾冁 - 傞 - 儲並儲 - 厘劊冗 - 乍佨乍 - 哋喒 - 値匒 - 唛偞劕 - 俈 - 俫喦 - 凝 - 喊唰 - 唾冗 - 二凧偷亍伡佒 - 勜亳 - 凶亦俞 - 介僆 - 刽仑 - 嗀 - 偣亮刚啺 - 丧唽 - 传喰先 - 呦匥冊 - 厑 - 協 - 僓偒傰 - 中侦佖 - 匿 - 參匉 - 剭丮劣丗 - 倎七 - 他喊匞 - 史冖否佥 - 啾 - 哘勤刐 - 叴伙劉仮 - 佧 - 劗 - 侒伿侒 - 亼 - 卒佁匄 - 刀凑 - 價傶價 - 刍 - 匊叅 - 仯佹 - 叛偶 - 冀唳 - 儲么儲 - 丁僁呧 - 叀 - 凝兓咽 - 区乢 - 儸唑 - 厙 - 卍 - 啿 - 儬咭仪 - 例 - 刜厶 - 义 - 呝 - 勘 - 僁呧叱 - 传匸 - 丗函 - 乣乩乣 - 儚乞兪 - 丁僁 - 伣 - 僬 - 儘 - 传喰匸 - 厮 - 剜 - 卛冶 - 啳丅 - 唏僊 - 卶咿交 - 僁亦 - 厥卶咿 - 伨倎七 - ▁仗勯剄 - 册 - 刉 - 個克僽 - 伞喱 - 呗 - 仦 - 啐 - 僜 - 冺 - 唃 - 呸 - 傂 - 劍 - 呀 - 亾 - 占 - 厯 - 介 - 偵 - 冾 - 啋 - 侾 - 咭 - 倃 - 劊 - 嗉 - 伩 - 击 - 倉 - 偊 - 伷 - 乴 - 参 - 亲 - 儡 - 亟 - 冇 - 分 - 呯 - 为 - 偁 - 咼 - 厠 - 切 - 健 - 凬 - 冠 - 伛 - 伥 - 叴 - 侼 - 余 - 偻 - 吥 - 佬 - 喏 - 卋 - 厼 - 品 - 倨 - 剡 - 喣 - 兵 - 勳 - 儨 - 匥 - 凷 - 乇 - 丛 - 咩 - 咐 - 傌 - 偨 - 匟 - 唂 - 匆 - 哱 - 兇 - 修 - 俷 - 叙 - 啜 - 啂 - 唉 - 侜 - 佽 - 喺 - 功 - 劝 - 剥 - 乎 - 咜 - 啥 - 便 - 哫 - 剓 - 劦 - 们 - 僠 - 刔 - 吟 - 唝 - 受 - 佺 - 凐 - 喻 - 儖 - 匧 - 勑 - 像 - 倦 - 僒 - 咉 - 僲 - 呦 - 剗 - 何 - 伭 - 侌 - 凋 - 側 - 侔 - 化 - 匘 - 典 - 処 - 吒 - 卙 - 傡 - 儜 - 剐 - 償 - 亚 - 呅 - 厇 - 啑 - 县 - 冼 - 冢 - 倭 - 僼 - 侄 - 卬 - 哤 - 哐 - 勯 - 倛 - 勌 - 剨 - 喢 - 勜 - 劵 - 乗 - 卡 - 伌 - 唀 - 咚 - 僶 - 勪 - 仐 - 哮 - 儍 - 冮 - 兘 - 偝 - 勉 - 俠 - 冫 - 傫 - 匚 - 伖 - 兰 - 凖 - 俙 - 世 - 叶 - 倫 - 匜 - 乹 - 咪 - 伙 - 之 - 啮 - 响 - 剀 - 卷 - 农 - 勧 - 勣 - 啖 - 倔 - 呲 - 刺 - 吔 - 啬 - 厨 - 勴 - 净 - 刴 - 叄 - 厣 - 厫 - 仂 - 傸 - 勆 - 凞 - 唈 - 刞 - 偺 - 俺 - 叕 - 匷 - 刖 - 凤 - 冕 - 劃 - 啄 - 冬 - 俀 - 唚 - 偐 - 儌 - 劉 - 俎 - 僖 - 刧 - 伱 - 劅 - 信 - 唕 - 伧 - 哇 - 协 - 倕 - 值 - 唐 - 佄 - 厕 - 冋 - 剢 - 吸 - 円 - 俳 - 于 - 哼 - 咨 - 吷 - 喅 - 呂 - 哔 - 咏 - 儬 - 凃 - 乨 - 吤 - 侭 - 匐 - 俊 - 倶 - 倱 - 匉 - 元 - 哺 - 儞 - 咸 - 単 - 厾 - 刲 - 佪 - 呴 - 傈 - 僚 - 丱 - 吓 - 叻 - 咖 - 冦 - 亱 - 呱 - 哻 - 卑 - 凌 - 兜 - 凜 - 嗏 - 倂 - 勍 - 剌 - 勅 - 勒 - 吊 - 厖 - 匌 - 卨 - 厈 - 儫 - 哦 - 伸 - 偩 - 僱 - 咫 - 偱 - 僥 - 吂 - 侗 - 亪 - 俔 - 匤 - 嗂 - 哗 - 吪 - 呮 - 吃 - 凱 - 卧 - 勦 - 侬 - 唻 - 劲 - 喯 - 卮 - 凳 - 友 - 啶 - 傁 - 喕 - 取 - 動 - 咯 - 劙 - 労 - 叾 - 剋 - 冲 - 各 - 偑 - 割 - 喟 - 咵 - 傔 - 发 - 啤 - 内 - 單 - 冥 - 吱 - 呩 - 務 - 咽 - 刮 - 刬 - 哌 - 叇 - 參 - 刵 - 咊 - 咰 - 冏 - 兿 - 吜 - 削 - 办 - 哃 - 喪 - 及 - 剺 - 劾 - 买 - 卩 - 卿 - 剔 - 势 - 呢 - 喠 - 俒 - 剹 - 偰 - 佰 - 匒 - 俕 - 劎 - 力 - 劶 - 叚 - 僺 - 兯 - 喁 - 咥 - 儽 - 呞 - 佮 - 创 - 医 - 低 - 俬 - 亂 - 党 - 匵 - 啙 - 咷 - 僅 - 僫 - 刣 - 咝 - 叆 - 变 - 勔 - 侁 - 倈 - 卝 - 候 - 傦 - 哶 - 啁 - 吴 - 吀 - 唙 - 卲 - 啀 - 儅 - 呤 - 儐 - 倗 - 兔 - 冐 - 亝 - 君 - 呕 - 僢 - 动 - 咒 - 卦 - 制 - 儢 - 劒 - 啠 - 仕 - 啔 - 咟 - 凯 - 唱 - 丬 - 哟 - 哂 - 冷 - 喈 - 劥 - 凕 - 偸 - 儒 - 侏 - 僂 - 喚 - 叨 - 冘 - 凩 - 侱 - 劭 - 卹 - 刃 - 兦 - 刈 - 啽 - 伆 - 嗆 - 唪 - 傿 - 呆 - 压 - 厡 - 厂 - 唔 - 仞 - 咴 - 僡 - 侍 - 儶 - 匣 - 劚 - 唨 - 乁 - 儵 - 卖 - 咳 - 偧 - 事 - 劻 - 刕 - 厷 - 佌 - 厽 - 呎 - 兤 - 呒 - 嗅 - 唠 - 剸 - 吝 - 卆 - 冿 - 啷 - 唞 - 凅 - 劁 - 剪 - 咻 - 厺 - 仄 - 哹 - 僸 - 勸 - 剉 - 劌 - 亁 - 呋 - 咄 - 哠 - 仔 - 副 - 凨 - 剾 - 傃 - 剶 - 侀 - 啛 - 勷 - 嗈 - 吩 - 唧 - 匲 - 厳 - 勏 - 凚 - 匝 - 儋 - 唲 - 剼 - 哑 - 匓 - 募 - 兂 - 匯 - 唺 - 咱 - 剳 - 嗄 - 嗋 - 乒 - 厝 - 喹 - 偉 - 剑 - 利 - 呟 - 冞 - 喞 - 哧 - 哰 - 劂 - 喾 - 傢 - 劐 - 匰 - 哷 - 傠 - 勂 - 儙 - 吳 - 啦 - 佷 - 剕 - 呺 - 吕 - 喖 - 吵 - 呼 - 喔 - 啻 - 六 - 吗 - 僪 - 凇 - 呄 - 冽 - 哝 - 咇 - 反 - 半 - 十 - 包 - 僞 - 兞 - 唯 - 呷 - 匽 - 儆 - 倒 - 剮 - 凈 - 劔 - 呰 - 刼 - 叟 - 劧 - 倸 - 哎 - 亯 - 叉 - 凭 - 减 - 可 - 兣 - 劷 - 倣 - 厏 - 偃 - 劀 - 乳 - 呜 - 司 - 嗁 - 傻 - 则 - 偆 - 匀 - 啪 - 呠 - 刪 - 喘 - 厁 - 喴 - 啼 - 啎 - 刱 - 咣 - 呚 - 卪 - 偕 - 允 - 努 - 劢 - 勄 - 啕 - 匇 - 傄 - 哳 - 勭 - 命 - 呙 - 唋 - 倯 - 啒 - 勹 - 剿 - 啘 - 侈 - 叁 - 俣 - 喫 - 呛 - 啈 - 原 - 吮 - 呑 - 哴 - 兴 - 俨 - 儹 - 厗 - 唡 - 伇 - 喼 - 卵 - 儛 - 咍 - 嗌 - 俭 - 凄 - 呾 - 冣 - 収 - 匑 - 别 - 僐 - 僿 - 印 - 卼 - 咲 - 哖 - 勞 - 倥 - 咹 - 儴 - 哆 - 具 - 俱 - 催 - 唭 - 嗊 - 哪 - 决 - 儦 - 喀 - 嗍 - 傼 - 书 - 咃 - 去 - 唓 - 啞 - 卽 - 哄 - 僉 - 叓 - 卄 - 咈 - 伵 - 啣 - 唤 - 仗 - 丕 - 严 - 咧 - <sos/eos> init: null input_size: null ctc_conf: dropout_rate: 0.0 ctc_type: builtin reduce: true ignore_nan_grad: null zero_infinity: true use_preprocessor: true token_type: bpe src_token_type: bpe bpemodel: data/token_list/tgt_bpe_unigram6000_ts_en/bpe.model src_bpemodel: data/token_list/src_bpe_unigram3000_rm_wavlm_large_21_km2000/bpe.model non_linguistic_symbols: null cleaner: null g2p: null tokenizer_encode_conf: null src_tokenizer_encode_conf: enable_sampling: true alpha: 0.4 nbest_size: -1 frontend: embed frontend_conf: embed_dim: 512 positional_dropout_rate: 0.1 specaug: specaug specaug_conf: apply_time_warp: false time_warp_window: 5 time_warp_mode: bicubic apply_freq_mask: false freq_mask_width_range: - 0 - 10 num_freq_mask: 0 apply_time_mask: true time_mask_width_ratio_range: - 0.0 - 0.05 num_time_mask: 10 preencoder: null preencoder_conf: {} encoder: e_branchformer encoder_conf: output_size: 256 attention_heads: 4 attention_layer_type: rel_selfattn pos_enc_layer_type: rel_pos rel_pos_type: latest cgmlp_linear_units: 1024 cgmlp_conv_kernel: 31 use_linear_after_conv: false gate_activation: identity num_blocks: 12 dropout_rate: 0.1 positional_dropout_rate: 0.1 attention_dropout_rate: 0.1 input_layer: conv1d2 layer_drop_rate: 0.0 linear_units: 1024 positionwise_layer_type: linear use_ffn: true macaron_ffn: true merge_conv_kernel: 31 postencoder: null postencoder_conf: {} decoder: transformer decoder_conf: attention_heads: 4 linear_units: 2048 num_blocks: 6 dropout_rate: 0.1 positional_dropout_rate: 0.1 self_attention_dropout_rate: 0.1 src_attention_dropout_rate: 0.1 layer_drop_rate: 0.0 model: discrete_asr model_conf: ctc_weight: 0.3 lsm_weight: 0.1 length_normalized_loss: false share_decoder_input_output_embed: false share_encoder_decoder_input_embed: false required: - output_dir - src_token_list - token_list version: '202310' distributed: false ``` </details> ### Citing ESPnet ```BibTex @inproceedings{watanabe2018espnet, author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai}, title={{ESPnet}: End-to-End Speech Processing Toolkit}, year={2018}, booktitle={Proceedings of Interspeech}, pages={2207--2211}, doi={10.21437/Interspeech.2018-1456}, url={http://dx.doi.org/10.21437/Interspeech.2018-1456} } ``` or arXiv: ```bibtex @misc{watanabe2018espnet, title={ESPnet: End-to-End Speech Processing Toolkit}, author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai}, year={2018}, eprint={1804.00015}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
[ "BEAR" ]
ostapeno/library-phi_2-v3-5-flan-clusters
ostapeno
null
[ "region:us" ]
"2024-01-20T23:46:28Z"
2024-01-20T23:46:38+00:00
0
0
--- {} --- Number of experts present in the library: 5 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_lora_embed_5clustersc1o5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,adversarial_qa_droberta_generate_question,true_case,stream_qed,cot_esnli,cot_sensemaking,trec_1_0_0,super_glue_wic_1_0_2,super_glue_record_1_0_2,yelp_polarity_reviews_0_2_0,cos_e_v1_11_rationale,natural_questions_open_1_0_0,web_questions_whats_the_answer,anli_r3_0_1_0,lambada_1_0_0,dream_generate_last_utterance,glue_cola_2_0_0,ag_news_subset_1_0_0,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cot_creak,math_dataset_algebra__linear_1d_1_0_0,web_questions_question_answer,stream_aqua,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,web_questions_short_general_knowledge_q,cnn_dailymail_3_4_0,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,snli_1_1_0,cos_e_v1_11_i_think,wiki_hop_original_explain_relation,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,web_questions_get_the_answer,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,cos_e_v1_11_explain_why_human,word_segment,anli_r2_0_1_0,race_high_Write_a_multi_choice_question_for_the_following_article,trivia_qa_rc_1_1_0,wmt16_translate_de_en_1_0_0,anli_r1_0_1_0,cot_ecqa,glue_stsb_2_0_0,cos_e_v1_11_aligned_with_common_sense,aeslc_1_0_0,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,dream_answer_to_dialogue,kilt_tasks_hotpotqa_formulate,para_crawl_enes,race_high_Write_a_multi_choice_question_options_given_,social_i_qa_I_was_wondering,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0,glue_mrpc_2_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_lora_embed_5clustersc0o5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,super_glue_rte_1_0_2,wiki_qa_found_on_google,app_reviews_categorize_rating_using_review,race_middle_Is_this_the_right_answer,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,wiki_qa_Is_This_True_,unified_qa_science_inst,super_glue_multirc_1_0_2,app_reviews_convert_to_rating,race_high_Is_this_the_right_answer,cot_strategyqa,cot_ecqa_ii,glue_qqp_2_0_0,quarel_do_not_use,wiki_qa_exercise,kilt_tasks_hotpotqa_complex_question,wiki_qa_automatic_system,cot_creak_ii,quarel_heres_a_story,qasc_is_correct_1,quarel_choose_between,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,glue_qnli_2_0_0,cot_sensemaking_ii,glue_mnli_2_0_0,super_glue_copa_1_0_2,social_i_qa_Check_if_a_random_answer_is_valid_or_not,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_Show_choices_and_generate_index,qasc_is_correct_2,quarel_testing_students,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,glue_wnli_2_0_0,cot_qasc,cot_strategyqa_ii,quarel_logic_test,stream_aqua_ii | lora | | phi2_joint_lora_embed_5clustersc3o5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,huggingface_xsum,wiqa_what_is_the_final_step_of_the_following_process,cot_gsm8k,wmt16_translate_ro_en_1_0_0,gem_dart_1_1_0,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,gem_common_gen_1_1_0,duorc_SelfRC_build_story_around_qa,gem_wiki_lingua_english_en_1_1_0,gigaword_1_2_0,gem_web_nlg_en_1_1_0,app_reviews_generate_review,wiki_bio_what_content,wiki_bio_who,gem_e2e_nlg_1_1_0,cot_esnli_ii,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | phi2_joint_lora_embed_5clustersc2o5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,cos_e_v1_11_question_option_description_text,ropes_background_new_situation_answer,social_i_qa_Show_choices_and_generate_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,ropes_plain_background_situation,race_high_Read_the_article_and_answer_the_question_no_option_,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,kilt_tasks_hotpotqa_final_exam,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,sciq_Multiple_Choice,race_high_Taking_a_test,wiqa_does_the_supposed_perturbation_have_an_effect,cos_e_v1_11_question_description_option_text,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,cos_e_v1_11_question_option_description_id,ropes_new_situation_background_answer,quail_description_context_question_answer_text,ropes_given_background_situation,quail_context_question_answer_description_text,ropes_prompt_bottom_hint_beginning,wiqa_effect_with_string_answer,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,qasc_qa_with_separated_facts_5,dream_baseline,ropes_prompt_beginning,quail_context_question_answer_description_id,quartz_having_read_above_passage,quail_context_description_question_answer_text,cos_e_v1_11_question_description_option_id,ropes_read_background_situation,qasc_qa_with_separated_facts_1,ropes_plain_bottom_hint,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,ropes_plain_no_background,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,sciq_Multiple_Choice_Closed_Book_,race_middle_Taking_a_test,quartz_use_info_from_paragraph_question,qasc_qa_with_separated_facts_4,cosmos_qa_1_0_0,quail_no_prompt_id,quartz_read_passage_below_choose,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_,quartz_paragraph_question_plain_concat,ropes_prompt_mix,ropes_background_situation_middle,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_lora_embed_5clustersc4o5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,wiki_hop_original_generate_object,quoref_Found_Context_Online,adversarial_qa_droberta_tell_what_it_is,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,kilt_tasks_hotpotqa_combining_facts,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,wiki_hop_original_choose_best_object_affirmative_3,quoref_Guess_Title_For_Context,quac_1_0_0,quoref_Answer_Test,wiki_hop_original_choose_best_object_interrogative_1,duorc_SelfRC_question_answering,adversarial_qa_droberta_based_on,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,wiki_hop_original_choose_best_object_affirmative_1,duorc_ParaphraseRC_extract_answer,adversarial_qa_dbert_answer_the_following_q,duorc_SelfRC_answer_question,wiki_hop_original_choose_best_object_interrogative_2,adversarial_qa_droberta_question_context_answer,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,squad_v2_0_3_0_0,web_questions_potential_correct_answer,wiki_hop_original_generate_subject,wiki_bio_guess_person,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,wiki_hop_original_choose_best_object_affirmative_2,duorc_SelfRC_movie_director,adversarial_qa_dbert_based_on,duorc_SelfRC_generate_question,wiki_hop_original_generate_subject_and_object,adversarial_qa_dbidaf_based_on,drop_2_0_0,adversarial_qa_dbert_question_context_answer,quoref_Given_Context_Answer_Question,adversarial_qa_dbidaf_tell_what_it_is,squad_v1_1_3_0_0 | lora | Last updated on: 2024-01-20 23:46:28+00:00
[ "SCIQ" ]
ostapeno/library-gptneo1B-10-flan-clusters
ostapeno
null
[ "region:us" ]
"2024-01-21T20:29:40Z"
2024-01-21T20:29:59+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | gptneo_jointc_lora_embed_clustersc0o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,duorc_ParaphraseRC_generate_question,adversarial_qa_dbidaf_generate_question,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,fix_punct,imdb_reviews_plain_text_1_0_0,wiki_hop_original_choose_best_object_affirmative_1,quail_context_description_question_answer_text,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_generate_subject,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,wiki_hop_original_choose_best_object_affirmative_2,cosmos_qa_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_missing_first_step,quail_no_prompt_text | lora | | gptneo_jointc_lora_embed_clustersc7o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,quoref_Find_Answer,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,adversarial_qa_droberta_tell_what_it_is,quoref_Read_And_Extract_,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,quoref_Guess_Title_For_Context,quoref_Answer_Test,duorc_SelfRC_question_answering,adversarial_qa_droberta_based_on,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,adversarial_qa_droberta_question_context_answer,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,adversarial_qa_dbert_tell_what_it_is,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,adversarial_qa_dbert_question_context_answer,quoref_Given_Context_Answer_Question,adversarial_qa_dbidaf_tell_what_it_is | lora | | gptneo_jointc_lora_embed_clustersc4o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0,dream_read_the_following_conversation_and_answer_the_question,race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,true_case,quail_description_context_question_answer_id,quail_context_question_description_text,stream_qed,cot_esnli,quoref_Context_Contains_Answer,race_high_Read_the_article_and_answer_the_question_no_option_,duorc_ParaphraseRC_movie_director,quail_context_description_question_answer_id,ag_news_subset_1_0_0,wiki_hop_original_generate_object,race_high_Taking_a_test,cos_e_v1_11_question_description_option_text,kilt_tasks_hotpotqa_combining_facts,race_middle_Select_the_best_answer,kilt_tasks_hotpotqa_straighforward_qa,quail_context_question_description_answer_id,quac_1_0_0,wiki_hop_original_explain_relation,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,dream_baseline,adversarial_qa_dbert_answer_the_following_q,quail_context_question_answer_description_id,app_reviews_generate_review,wiki_bio_what_content,quoref_Answer_Question_Given_Context,trivia_qa_rc_1_1_0,wiki_bio_guess_person,glue_stsb_2_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,gem_e2e_nlg_1_1_0,race_middle_Taking_a_test,duorc_SelfRC_movie_director,quail_no_prompt_id,wiki_hop_original_generate_subject_and_object,race_middle_Select_the_best_answer_no_instructions_,squad_v1_1_3_0_0 | lora | | gptneo_jointc_lora_embed_clustersc3o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,wiqa_what_is_the_final_step_of_the_following_process,yelp_polarity_reviews_0_2_0,cos_e_v1_11_rationale,anli_r3_0_1_0,duorc_SelfRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,gem_dart_1_1_0,cos_e_v1_11_generate_explanation_given_text,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,duorc_ParaphraseRC_title_generation,cos_e_v1_11_i_think,gem_wiki_lingua_english_en_1_1_0,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,gigaword_1_2_0,kilt_tasks_hotpotqa_complex_question,gem_web_nlg_en_1_1_0,cos_e_v1_11_explain_why_human,word_segment,anli_r2_0_1_0,wmt16_translate_de_en_1_0_0,anli_r1_0_1_0,cos_e_v1_11_aligned_with_common_sense,aeslc_1_0_0,wmt16_translate_fi_en_1_0_0,race_high_Write_a_multi_choice_question_options_given_,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0,cot_esnli_ii | lora | | gptneo_jointc_lora_embed_clustersc9o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,ropes_plain_background_situation,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,quartz_use_info_from_question_paragraph,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,quartz_use_info_from_paragraph_question,quartz_paragraph_question_plain_concat,ropes_prompt_mix,ropes_background_situation_middle | lora | | gptneo_jointc_lora_embed_clustersc5o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2,trec_1_0_0,natural_questions_open_1_0_0,lambada_1_0_0,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,kilt_tasks_hotpotqa_final_exam,glue_cola_2_0_0,paws_wiki_1_1_0,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,wiki_qa_Is_This_True_,unified_qa_science_inst,stream_aqua,super_glue_multirc_1_0_2,snli_1_1_0,cot_strategyqa,cot_ecqa_ii,glue_qqp_2_0_0,cot_creak_ii,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,glue_qnli_2_0_0,cot_sensemaking_ii,glue_mnli_2_0_0,super_glue_copa_1_0_2,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,glue_wnli_2_0_0,glue_mrpc_2_0_0,cot_qasc,cot_strategyqa_ii,stream_aqua_ii | lora | | gptneo_jointc_lora_embed_clustersc1o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,cos_e_v1_11_question_option_description_text,wiki_qa_found_on_google,social_i_qa_Show_choices_and_generate_answer,app_reviews_categorize_rating_using_review,super_glue_wic_1_0_2,super_glue_record_1_0_2,race_middle_Is_this_the_right_answer,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,race_high_Select_the_best_answer_no_instructions_,sciq_Multiple_Choice,wiqa_does_the_supposed_perturbation_have_an_effect,math_dataset_algebra__linear_1d_1_0_0,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cos_e_v1_11_question_option_description_id,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,race_high_Is_this_the_right_answer,wiqa_effect_with_string_answer,quarel_do_not_use,qasc_qa_with_separated_facts_5,wiki_qa_exercise,ropes_prompt_beginning,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,qasc_qa_with_separated_facts_1,wiki_qa_automatic_system,cos_e_v1_11_description_question_option_text,quarel_heres_a_story,qasc_qa_with_combined_facts_1,qasc_is_correct_1,squad_v2_0_3_0_0,cos_e_v1_11_description_question_option_id,quarel_choose_between,social_i_qa_Check_if_a_random_answer_is_valid_or_not,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,social_i_qa_I_was_wondering,qasc_is_correct_2,quarel_testing_students,qasc_qa_with_separated_facts_4,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,quartz_read_passage_below_choose,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,drop_2_0_0,social_i_qa_Generate_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | gptneo_jointc_lora_embed_clustersc6o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/web_questions_whats_the_answer,web_questions_question_answer,web_questions_short_general_knowledge_q,web_questions_get_the_answer,web_questions_potential_correct_answer | lora | | gptneo_jointc_lora_embed_clustersc8o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/huggingface_xsum,cot_gsm8k,cot_sensemaking,cot_creak,cnn_dailymail_3_4_0,race_middle_Write_a_multi_choice_question_for_the_following_article,race_high_Write_a_multi_choice_question_for_the_following_article,dream_generate_first_utterance,para_crawl_enes,wmt16_translate_tr_en_1_0_0 | lora | | gptneo_jointc_lora_embed_clustersc2o10_2e_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_generate_last_utterance,wiki_bio_key_content,duorc_SelfRC_build_story_around_qa,cot_ecqa,wiki_bio_who,dream_answer_to_dialogue,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | Last updated on: 2024-01-21 20:29:41+00:00
[ "SCIQ" ]
sordonia/library-t5_xl_lm-2epc
sordonia
null
[ "region:us" ]
"2024-01-22T15:03:58Z"
2024-01-25T00:53:55+00:00
0
0
--- {} --- Number of experts present in the library: 38 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | wiqa | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/wiqa | lora | | amazon_polarity | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/amazon_polarity | lora | | quartz | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/quartz | lora | | rotten_tomatoes | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/rotten_tomatoes | lora | | xsum | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/xsum | lora | | wiki_bio | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/wiki_bio | lora | | common_gen | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/common_gen | lora | | wiki_hop_original | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/wiki_hop_original | lora | | glue_mrpc | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/glue_mrpc | lora | | duorc_SelfRC | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/duorc_SelfRC | lora | | trec | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/trec | lora | | paws_labeled_final | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/paws_labeled_final | lora | | quarel | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/quarel | lora | | ag_news | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/ag_news | lora | | glue_qqp | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/glue_qqp | lora | | quail | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/quail | lora | | cos_e_v1_11 | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/cos_e_v1_11 | lora | | imdb | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/imdb | lora | | adversarial_qa_dbert | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/adversarial_qa_dbert | lora | | cosmos_qa | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/cosmos_qa | lora | | qasc | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/qasc | lora | | adversarial_qa_droberta | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/adversarial_qa_droberta | lora | | quoref | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/quoref | lora | | social_i_qa | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/social_i_qa | lora | | app_reviews | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/app_reviews | lora | | cnn_dailymail_3_0_0 | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/cnn_dailymail_3_0_0 | lora | | adversarial_qa_dbidaf | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/adversarial_qa_dbidaf | lora | | dream | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/dream | lora | | dbpedia_14 | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/dbpedia_14 | lora | | wiki_qa | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/wiki_qa | lora | | samsum | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/samsum | lora | | multi_news | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/multi_news | lora | | ropes | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/ropes | lora | | duorc_ParaphraseRC | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/duorc_ParaphraseRC | lora | | kilt_tasks_hotpotqa | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/kilt_tasks_hotpotqa | lora | | gigaword | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/gigaword | lora | | sciq | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/sciq | lora | | yelp_review_full | google/t5-xl-lm-adapt | sordonia/t0-1.6M-flat/yelp_review_full | lora | Last updated on: 2024-01-22 15:04:02+00:00
[ "SCIQ" ]
zhan1993/library-phi_2-v3-10-clusters-loss
zhan1993
null
[ "region:us" ]
"2024-01-23T04:15:31Z"
2024-01-23T04:15:39+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_3epoch_loss_cluster_7 | phi-2 | sordonia/flan-10k-flat/wiki_bio_comprehension,wiki_bio_what_content,cos_e_v1_11_aligned_with_common_sense,gem_web_nlg_en_1_1_0,cnn_dailymail_3_4_0,race_high_Write_a_multi_choice_question_options_given_,stream_aqua_ii,duorc_SelfRC_build_story_around_qa,cot_esnli_ii,cos_e_v1_11_i_think,stream_qed_ii,cos_e_v1_11_rationale,race_middle_Write_a_multi_choice_question_options_given_,trivia_qa_rc_1_1_0,cot_sensemaking_ii,wiki_bio_key_content,dream_generate_first_utterance,duorc_ParaphraseRC_build_story_around_qa,app_reviews_convert_to_rating,wiqa_what_is_the_final_step_of_the_following_process,cot_strategyqa,wiki_qa_Direct_Answer_to_Question,wiqa_what_might_be_the_first_step_of_the_process,race_middle_Write_a_multi_choice_question_for_the_following_article,wiqa_what_might_be_the_last_step_of_the_process,cos_e_v1_11_generate_explanation_given_text,true_case,cos_e_v1_11_explain_why_human,race_high_Write_a_multi_choice_question_for_the_following_article,huggingface_xsum,cot_creak_ii,wiki_bio_who,dream_answer_to_dialogue,word_segment | lora | | phi2_joint_3epoch_loss_cluster_9 | phi-2 | sordonia/flan-10k-flat/stream_aqua,cot_esnli,fix_punct,cot_qasc,cot_gsm8k,multi_news_1_0_0,cot_creak,stream_qed,cot_sensemaking,cot_ecqa | lora | | phi2_joint_3epoch_loss_cluster_8 | phi-2 | sordonia/flan-10k-flat/trec_1_0_0,glue_qqp_2_0_0,paws_wiki_1_1_0,anli_r3_0_1_0,anli_r2_0_1_0,anli_r1_0_1_0,ag_news_subset_1_0_0 | lora | | phi2_joint_3epoch_loss_cluster_10 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,imdb_reviews_plain_text_1_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,snli_1_1_0,super_glue_multirc_1_0_2,super_glue_wic_1_0_2,definite_pronoun_resolution_1_1_0,super_glue_rte_1_0_2,glue_mrpc_2_0_0,glue_wnli_2_0_0,super_glue_wsc_fixed_1_0_2,glue_cola_2_0_0 | lora | | phi2_joint_3epoch_loss_cluster_1 | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice,qasc_qa_with_separated_facts_1,qasc_qa_with_separated_facts_2,quartz_having_read_above_passage,dream_baseline,sciq_Multiple_Choice_Closed_Book_,qasc_qa_with_separated_facts_4,quail_context_question_answer_description_text,sciq_Multiple_Choice_Question_First,dream_read_the_following_conversation_and_answer_the_question,qasc_qa_with_combined_facts_1 | lora | | phi2_joint_3epoch_loss_cluster_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1,quarel_testing_students,quartz_use_info_from_paragraph_question,ropes_plain_bottom_hint,ropes_prompt_mix,ropes_read_background_situation,wiki_hop_original_choose_best_object_affirmative_1,quail_no_prompt_text,ropes_given_background_situation,quarel_logic_test,quail_context_question_description_answer_text,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_affirmative_2,wiqa_effect_with_string_answer,quail_context_description_question_answer_text,quartz_read_passage_below_choose,wiki_hop_original_choose_best_object_affirmative_3,quarel_choose_between,sciq_Direct_Question_Closed_Book_,quartz_answer_question_below,ropes_prompt_bottom_no_hint,quartz_given_the_fact_answer_the_q,race_middle_Select_the_best_answer_generate_span_,ropes_background_new_situation_answer,sciq_Direct_Question,social_i_qa_Show_choices_and_generate_answer,quartz_answer_question_based_on,quarel_do_not_use,ropes_prompt_beginning,quarel_heres_a_story,ropes_background_situation_middle,ropes_new_situation_background_answer,ropes_plain_background_situation,quail_description_context_question_answer_text,quartz_paragraph_question_plain_concat,wiqa_which_of_the_following_is_the_supposed_perturbation,quartz_use_info_from_question_paragraph,ropes_plain_no_background,race_high_Select_the_best_answer_generate_span_,ropes_prompt_bottom_hint_beginning | lora | | phi2_joint_3epoch_loss_cluster_4 | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise,duorc_SelfRC_generate_question_by_answer,cot_gsm8k_ii,gem_wiki_lingua_english_en_1_1_0,wiki_hop_original_generate_object,wiqa_what_is_the_missing_first_step,wiki_qa_found_on_google,wmt16_translate_de_en_1_0_0,quoref_Guess_Title_For_Context,gem_common_gen_1_1_0,super_glue_cb_1_0_2,para_crawl_enes,app_reviews_generate_review,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,super_glue_copa_1_0_2,math_dataset_algebra__linear_1d_1_0_0,app_reviews_convert_to_star_rating,wiki_bio_guess_person,wiqa_does_the_supposed_perturbation_have_an_effect,wiki_qa_Jeopardy_style,wmt16_translate_tr_en_1_0_0,duorc_ParaphraseRC_title_generation,race_middle_Is_this_the_right_answer,app_reviews_categorize_rating_using_review,gem_dart_1_1_0,duorc_ParaphraseRC_generate_question_by_answer,wiki_hop_original_generate_subject,aeslc_1_0_0,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,wmt16_translate_ro_en_1_0_0,race_high_Is_this_the_right_answer,duorc_SelfRC_title_generation,duorc_SelfRC_generate_question,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_hop_original_generate_subject_and_object,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_explain_relation,super_glue_record_1_0_2,cot_ecqa_ii,wmt14_translate_fr_en_1_0_0,dbpedia_14_given_a_choice_of_categories_,wiki_qa_Generate_Question_from_Topic,dbpedia_14_pick_one_category_for_the_following_text,gem_e2e_nlg_1_1_0,social_i_qa_Generate_the_question_from_the_answer,cot_strategyqa_ii,wmt16_translate_fi_en_1_0_0,duorc_ParaphraseRC_generate_question,dream_generate_last_utterance,quac_1_0_0 | lora | | phi2_joint_3epoch_loss_cluster_3 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0,adversarial_qa_dbert_generate_question,quoref_Found_Context_Online,web_questions_get_the_answer,unified_qa_science_inst,quoref_What_Is_The_Answer,adversarial_qa_droberta_generate_question,adversarial_qa_dbidaf_question_context_answer,web_questions_whats_the_answer,adversarial_qa_droberta_question_context_answer,kilt_tasks_hotpotqa_combining_facts,web_questions_potential_correct_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_decide_worth_it,adversarial_qa_droberta_tell_what_it_is,duorc_ParaphraseRC_answer_question,duorc_SelfRC_extract_answer,drop_2_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_dbidaf_based_on,natural_questions_open_1_0_0,lambada_1_0_0,quoref_Read_And_Extract_,duorc_SelfRC_question_answering,squad_v1_1_3_0_0,coqa_1_0_0,web_questions_question_answer,wiki_qa_Topic_Prediction_Question_Only,quoref_Find_Answer,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,cos_e_v1_11_question_description_option_text,duorc_SelfRC_movie_director,quoref_Given_Context_Answer_Question,adversarial_qa_dbidaf_tell_what_it_is,web_questions_short_general_knowledge_q,duorc_SelfRC_answer_question,quoref_Answer_Question_Given_Context,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_text,adversarial_qa_dbert_based_on,duorc_ParaphraseRC_movie_director,quoref_Guess_Answer,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,kilt_tasks_hotpotqa_final_exam,quoref_Answer_Friend_Question,adversarial_qa_dbert_question_context_answer,kilt_tasks_hotpotqa_complex_question,wiki_qa_Topic_Prediction_Answer_Only,duorc_ParaphraseRC_decide_worth_it,quoref_Answer_Test,kilt_tasks_hotpotqa_formulate,gigaword_1_2_0,duorc_ParaphraseRC_question_answering,adversarial_qa_droberta_based_on | lora | | phi2_joint_3epoch_loss_cluster_5 | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test,quail_no_prompt_id,quail_context_question_answer_description_id,race_high_Select_the_best_answer,race_high_Select_the_best_answer_no_instructions_,wiqa_effect_with_label_answer,race_middle_Select_the_best_answer,cos_e_v1_11_question_description_option_id,quail_context_description_question_answer_id,quail_description_context_question_answer_id,wiki_qa_Decide_good_answer,quail_context_question_description_answer_id,race_middle_Select_the_best_answer_no_instructions_,cos_e_v1_11_description_question_option_id,race_middle_Taking_a_test,cos_e_v1_11_question_option_description_id,social_i_qa_Show_choices_and_generate_index | lora | | phi2_joint_3epoch_loss_cluster_6 | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_,race_high_Read_the_article_and_answer_the_question_no_option_,yelp_polarity_reviews_0_2_0,qasc_is_correct_1,social_i_qa_Generate_answer,cosmos_qa_1_0_0,qasc_qa_with_separated_facts_5,glue_stsb_2_0_0,social_i_qa_I_was_wondering,qasc_qa_with_separated_facts_3,quail_description_context_question_text,qasc_is_correct_2,quail_context_question_description_text,social_i_qa_Check_if_a_random_answer_is_valid_or_not,race_middle_Read_the_article_and_answer_the_question_no_option_,quail_context_description_question_text,wiki_qa_automatic_system | lora | Last updated on: 2024-01-23T04:15:32.000Z
[ "SCIQ" ]
ntc-ai/SDXL-LoRA-slider.smoking-a-cigarette-looking-cool
ntc-ai
text-to-image
[ "diffusers", "text-to-image", "stable-diffusion-xl", "lora", "template:sd-lora", "template:sdxl-lora", "sdxl-sliders", "ntcai.xyz-sliders", "concept", "en", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:mit", "region:us" ]
"2024-01-23T13:25:41Z"
2024-01-23T13:25:44+00:00
0
0
--- base_model: stabilityai/stable-diffusion-xl-base-1.0 language: - en license: mit tags: - text-to-image - stable-diffusion-xl - lora - template:sd-lora - template:sdxl-lora - sdxl-sliders - ntcai.xyz-sliders - concept - diffusers thumbnail: images/evaluate/smoking a cigarette looking cool.../smoking a cigarette looking cool_17_3.0.png widget: - text: smoking a cigarette looking cool output: url: images/smoking a cigarette looking cool_17_3.0.png - text: smoking a cigarette looking cool output: url: images/smoking a cigarette looking cool_19_3.0.png - text: smoking a cigarette looking cool output: url: images/smoking a cigarette looking cool_20_3.0.png - text: smoking a cigarette looking cool output: url: images/smoking a cigarette looking cool_21_3.0.png - text: smoking a cigarette looking cool output: url: images/smoking a cigarette looking cool_22_3.0.png inference: false instance_prompt: smoking a cigarette looking cool --- # ntcai.xyz slider - smoking a cigarette looking cool (SDXL LoRA) | Strength: -3 | Strength: 0 | Strength: 3 | | --- | --- | --- | | <img src="images/smoking a cigarette looking cool_17_-3.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_17_0.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_17_3.0.png" width=256 height=256 /> | | <img src="images/smoking a cigarette looking cool_19_-3.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_19_0.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_19_3.0.png" width=256 height=256 /> | | <img src="images/smoking a cigarette looking cool_20_-3.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_20_0.0.png" width=256 height=256 /> | <img src="images/smoking a cigarette looking cool_20_3.0.png" width=256 height=256 /> | ## Download Weights for this model are available in Safetensors format. ## Trigger words You can apply this LoRA with trigger words for additional effect: ``` smoking a cigarette looking cool ``` ## Use in diffusers ```python from diffusers import StableDiffusionXLPipeline from diffusers import EulerAncestralDiscreteScheduler import torch pipe = StableDiffusionXLPipeline.from_single_file("https://huggingface.co/martyn/sdxl-turbo-mario-merge-top-rated/blob/main/topRatedTurboxlLCM_v10.safetensors") pipe.to("cuda") pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) # Load the LoRA pipe.load_lora_weights('ntc-ai/SDXL-LoRA-slider.smoking-a-cigarette-looking-cool', weight_name='smoking a cigarette looking cool.safetensors', adapter_name="smoking a cigarette looking cool") # Activate the LoRA pipe.set_adapters(["smoking a cigarette looking cool"], adapter_weights=[2.0]) prompt = "medieval rich kingpin sitting in a tavern, smoking a cigarette looking cool" negative_prompt = "nsfw" width = 512 height = 512 num_inference_steps = 10 guidance_scale = 2 image = pipe(prompt, negative_prompt=negative_prompt, width=width, height=height, guidance_scale=guidance_scale, num_inference_steps=num_inference_steps).images[0] image.save('result.png') ``` ## Support the Patreon If you like this model please consider [joining our Patreon](https://www.patreon.com/NTCAI). By joining our Patreon, you'll gain access to an ever-growing library of over 1140+ unique and diverse LoRAs, covering a wide range of styles and genres. You'll also receive early access to new models and updates, exclusive behind-the-scenes content, and the powerful LoRA slider creator, allowing you to craft your own custom LoRAs and experiment with endless possibilities. Your support on Patreon will allow us to continue developing and refining new models. ## Other resources - [CivitAI](https://civitai.com/user/ntc) - Follow ntc on Civit for even more LoRAs - [ntcai.xyz](https://ntcai.xyz) - See ntcai.xyz to find more articles and LoRAs
[ "CRAFT" ]
ostapeno/library-gptneo1B-10-rand-flan-clusters
ostapeno
null
[ "region:us" ]
"2024-01-23T17:00:30Z"
2024-09-18T19:26:25+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | gptneo_jointc_lora_embed_randclusterscluster_3_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5,wiki_qa_automatic_system,stream_aqua_ii,dbpedia_14_pick_one_category_for_the_following_text,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Generate_Question_from_Topic,cot_gsm8k_ii,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_interrogative_2,quail_context_question_answer_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,qasc_qa_with_combined_facts_1,adversarial_qa_dbert_answer_the_following_q,social_i_qa_I_was_wondering,stream_aqua,word_segment,ropes_plain_no_background,super_glue_multirc_1_0_2,wiki_hop_original_choose_best_object_affirmative_3,app_reviews_convert_to_rating,anli_r3_0_1_0,app_reviews_convert_to_star_rating,quartz_paragraph_question_plain_concat,kilt_tasks_hotpotqa_complex_question,quartz_use_info_from_paragraph_question | lora | | gptneo_jointc_lora_embed_randclusterscluster_9_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article,cot_qasc,definite_pronoun_resolution_1_1_0,wiki_qa_found_on_google,wiki_bio_comprehension,wiki_qa_Topic_Prediction_Question_Only,wiki_bio_guess_person,fix_punct,race_middle_Select_the_best_answer_no_instructions_,quac_1_0_0,wiqa_does_the_supposed_perturbation_have_an_effect,quartz_given_the_fact_answer_the_q,ropes_new_situation_background_answer,social_i_qa_Generate_answer,gigaword_1_2_0,duorc_SelfRC_decide_worth_it,kilt_tasks_hotpotqa_straighforward_qa,quail_no_prompt_text,cot_esnli,quoref_Answer_Friend_Question,race_middle_Select_the_best_answer_generate_span_,unified_qa_science_inst,sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_strategyqa | lora | | gptneo_jointc_lora_embed_randclusterscluster_5_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0,duorc_SelfRC_question_answering,cot_sensemaking_ii,huggingface_xsum,duorc_ParaphraseRC_build_story_around_qa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,quartz_answer_question_based_on,ropes_plain_background_situation,race_middle_Write_a_multi_choice_question_for_the_following_article,quail_description_context_question_text,web_questions_whats_the_answer,cot_ecqa,true_case,adversarial_qa_dbert_question_context_answer,duorc_SelfRC_title_generation,quail_context_question_description_answer_id,quarel_do_not_use,adversarial_qa_dbidaf_answer_the_following_q,duorc_ParaphraseRC_decide_worth_it,race_middle_Is_this_the_right_answer,wmt16_translate_de_en_1_0_0,wiki_qa_Is_This_True_,race_middle_Select_the_best_answer,aeslc_1_0_0,duorc_ParaphraseRC_question_answering,wiki_hop_original_choose_best_object_affirmative_1 | lora | | gptneo_jointc_lora_embed_randclusterscluster_4_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_explain_why_human,cot_sensemaking,quoref_Find_Answer,quail_context_description_question_text,social_i_qa_Generate_the_question_from_the_answer,quartz_having_read_above_passage,stream_qed_ii,wiqa_what_might_be_the_first_step_of_the_process,wiki_bio_who,duorc_SelfRC_generate_question_by_answer,race_high_Select_the_best_answer_generate_span_,lambada_1_0_0,coqa_1_0_0,race_high_Select_the_best_answer,adversarial_qa_droberta_question_context_answer,quail_context_question_answer_description_id,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_tell_what_it_is,cnn_dailymail_3_4_0,quail_no_prompt_id,ag_news_subset_1_0_0,trivia_qa_rc_1_1_0,ropes_prompt_bottom_hint_beginning,super_glue_wic_1_0_2 | lora | | gptneo_jointc_lora_embed_randclusterscluster_6_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0,para_crawl_enes,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_id,glue_stsb_2_0_0,quail_context_description_question_answer_text,dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_description_option_text,cos_e_v1_11_description_question_option_text,race_high_Write_a_multi_choice_question_options_given_,ropes_prompt_bottom_no_hint,quarel_testing_students,wmt16_translate_tr_en_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_final_step_of_the_following_process,quoref_Given_Context_Answer_Question,wiki_hop_original_generate_object,quartz_use_info_from_question_paragraph,duorc_SelfRC_build_story_around_qa,drop_2_0_0,wiqa_effect_with_string_answer,race_high_Taking_a_test,wiki_hop_original_generate_subject_and_object,glue_qqp_2_0_0,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | gptneo_jointc_lora_embed_randclusterscluster_1_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0,quarel_choose_between,wiki_hop_original_explain_relation,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,race_high_Select_the_best_answer_no_instructions_,kilt_tasks_hotpotqa_combining_facts,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_generate_question,glue_cola_2_0_0,imdb_reviews_plain_text_1_0_0,squad_v1_1_3_0_0,race_high_Is_this_the_right_answer,qasc_qa_with_separated_facts_1,glue_sst2_2_0_0,wiqa_what_is_the_missing_first_step,duorc_ParaphraseRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,wiki_qa_Jeopardy_style,quartz_answer_question_below,ropes_prompt_beginning,ropes_read_background_situation,wiki_qa_Decide_good_answer,super_glue_cb_1_0_2,qasc_qa_with_separated_facts_4,cot_ecqa_ii | lora | | gptneo_jointc_lora_embed_randclusterscluster_10_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed,cot_esnli_ii,quarel_heres_a_story,quoref_Guess_Title_For_Context,qasc_is_correct_2,wiqa_effect_with_label_answer,dream_generate_last_utterance,adversarial_qa_dbert_based_on,dream_answer_to_dialogue,sciq_Multiple_Choice_Question_First,quail_context_question_description_answer_text,wiki_qa_Direct_Answer_to_Question,ropes_background_new_situation_answer,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,quoref_What_Is_The_Answer,dbpedia_14_given_a_choice_of_categories_,qasc_qa_with_separated_facts_2,glue_mrpc_2_0_0,gem_e2e_nlg_1_1_0,anli_r1_0_1_0,race_middle_Read_the_article_and_answer_the_question_no_option_,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,adversarial_qa_dbidaf_based_on | lora | | gptneo_jointc_lora_embed_randclusterscluster_2_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id,social_i_qa_Show_choices_and_generate_index,web_questions_potential_correct_answer,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cosmos_qa_1_0_0,sciq_Direct_Question,super_glue_wsc_fixed_1_0_2,race_middle_Taking_a_test,wmt14_translate_fr_en_1_0_0,duorc_SelfRC_extract_answer,wiki_hop_original_generate_subject,duorc_SelfRC_answer_question,qasc_is_correct_1,cos_e_v1_11_i_think,wiki_qa_exercise,race_middle_Write_a_multi_choice_question_options_given_,quoref_Read_And_Extract_,web_questions_short_general_knowledge_q,web_questions_question_answer,quarel_logic_test,app_reviews_categorize_rating_using_review,cot_strategyqa_ii,glue_mnli_2_0_0,quoref_Answer_Test,super_glue_rte_1_0_2 | lora | | gptneo_jointc_lora_embed_randclusterscluster_7_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii,quail_description_context_question_answer_id,kilt_tasks_hotpotqa_final_exam,cot_gsm8k,cos_e_v1_11_aligned_with_common_sense,squad_v2_0_3_0_0,duorc_SelfRC_movie_director,anli_r2_0_1_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,duorc_ParaphraseRC_answer_question,ropes_background_situation_middle,dream_generate_first_utterance,quail_context_question_description_text,adversarial_qa_droberta_based_on,glue_wnli_2_0_0,super_glue_record_1_0_2,web_questions_get_the_answer,ropes_prompt_mix,app_reviews_generate_review,cos_e_v1_11_rationale,adversarial_qa_droberta_generate_question,yelp_polarity_reviews_0_2_0,ropes_given_background_situation,qasc_qa_with_separated_facts_3,quoref_Guess_Answer | lora | | gptneo_jointc_lora_embed_randclusterscluster_8_3epoch | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2,duorc_ParaphraseRC_movie_director,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,quartz_read_passage_below_choose,social_i_qa_Check_if_a_random_answer_is_valid_or_not,cot_creak,ropes_plain_bottom_hint,super_glue_copa_1_0_2,natural_questions_open_1_0_0,trec_1_0_0,gem_web_nlg_en_1_1_0,wiki_bio_key_content,wmt16_translate_fi_en_1_0_0,quoref_Answer_Question_Given_Context,duorc_ParaphraseRC_generate_question,math_dataset_algebra__linear_1d_1_0_0,duorc_ParaphraseRC_title_generation,quail_context_description_question_answer_id,wiki_bio_what_content,adversarial_qa_dbert_tell_what_it_is,sciq_Multiple_Choice_Closed_Book_,duorc_ParaphraseRC_extract_answer,dream_baseline,gem_wiki_lingua_english_en_1_1_0 | lora | Last updated on: 2024-01-23 17:00:30+00:00
[ "SCIQ" ]
sethuiyer/txtai-medsearch
sethuiyer
sentence-similarity
[ "txtai", "sentence-similarity", "en", "zh", "license:cc-by-sa-3.0", "license:gfdl", "region:us" ]
"2024-01-24T07:49:41Z"
2024-01-24T08:33:47+00:00
0
6
--- language: - en - zh library_name: txtai license: - cc-by-sa-3.0 - gfdl tags: - sentence-similarity inference: false --- # Medical txtai embeddings index This is a [txtai](https://github.com/neuml/txtai) embeddings index specifically designed for medical texts, encompassing a diverse corpus in both English and Chinese. The model is primed for integration into medical information systems, aiding in the quick retrieval of relevant clinical information. ### Data Sources The model is trained on a substantial dataset, including 434411 entries from a bilingual (English and Chinese) corpus of clinical texts. The sources are: - `shibing624/medical`, a dataset featuring a variety of medical scenarios and questions in both English and Chinese, suitable for text generation and medical question-answering systems. It's licensed under Apache 2.0. - `keivalya/MedQuad-MedicalQnADataset`, offering detailed insights into various health conditions and their treatments, covering prevention, diagnosis, treatment, and susceptibility. - `GBaker/MedQA-USMLE-4-options`, a collection of multiple-choice questions based on the USMLE, focusing on a wide range of medical topics and scenarios. - `medalpaca/medical_meadow_medqa`, a dataset for question answering in English and Chinese, encompassing clinical scenarios and medical queries with multiple-choice answers. - `medalpaca/medical_meadow_medical_flashcards`, featuring over 34,000 rows of question and answer pairs derived from medical flashcards, focusing on a wide range of medical subjects. Each of these datasets contributes to the depth and diversity of the medical knowledge encapsulated in the txtai embeddings model, making it an effective tool for medical information retrieval and analysis. ## Indexing The txtai embeddings model utilizes 'efederici/multilingual-e5-small-4096', a transformer-based model with 12 layers and an embedding size of 384, supporting 94 languages. ### Configuration The embedding model is quantized to 4 bits for size efficiency and supports batch encoding of 15 for optimized performance. The indexing is implemented using simple numpy cosine similarity, ensuring straightforward and efficient retrieval. ## Usage 1. Load the dataset using the provided JSON file. 2. Initialize and load the embeddings using txtai: ```python from txtai import Embeddings embeddings = Embeddings() embeddings.load('index.tar.gz') ``` ## Next Steps 1. More detailed usage, including using txtai to create inter-operability between English and Chinese 2. Create an usecase with [CrewAI](https://github.com/joaomdmoura/crewAI) and [Dr.Samantha](https://huggingface.co/sethuiyer/Dr_Samantha_7b_mistral) ## License This model is available under the Creative Commons Attribution-ShareAlike 3.0 Unported License and the GNU Free Documentation License.
[ "MEDQA" ]
zhan1993/library-phi_2-v3-10-clusters-sim-alex
zhan1993
null
[ "region:us" ]
"2024-01-24T13:39:11Z"
2024-07-06T14:03:15+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_3epoch_sim_cluster_10 | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,app_reviews_convert_to_star_rating,cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,kilt_tasks_hotpotqa_final_exam,sciq_Multiple_Choice,wiqa_does_the_supposed_perturbation_have_an_effect,cos_e_v1_11_question_description_option_text,wiki_qa_Is_This_True_,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cos_e_v1_11_question_option_description_id,wiqa_effect_with_string_answer,qasc_qa_with_separated_facts_5,dream_baseline,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,qasc_is_correct_1,cos_e_v1_11_description_question_option_id,social_i_qa_Check_if_a_random_answer_is_valid_or_not,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,qasc_is_correct_2,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_3epoch_sim_cluster_3 | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google,app_reviews_categorize_rating_using_review,race_middle_Is_this_the_right_answer,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,unified_qa_science_inst,race_high_Is_this_the_right_answer,cot_strategyqa,cot_ecqa_ii,quarel_do_not_use,wiki_qa_exercise,wiki_qa_automatic_system,cot_creak_ii,quarel_heres_a_story,quarel_choose_between,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,glue_qnli_2_0_0,cot_sensemaking_ii,super_glue_copa_1_0_2,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_Show_choices_and_generate_index,quarel_testing_students,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,glue_wnli_2_0_0,cot_qasc,cot_strategyqa_ii,quarel_logic_test,stream_aqua_ii | lora | | phi2_joint_3epoch_sim_cluster_9 | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2,cot_sensemaking,super_glue_wic_1_0_2,cos_e_v1_11_rationale,anli_r3_0_1_0,dream_generate_last_utterance,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cot_creak,stream_aqua,snli_1_1_0,cos_e_v1_11_i_think,glue_qqp_2_0_0,cos_e_v1_11_explain_why_human,anli_r2_0_1_0,anli_r1_0_1_0,glue_stsb_2_0_0,cos_e_v1_11_aligned_with_common_sense,glue_mnli_2_0_0,social_i_qa_I_was_wondering,cosmos_qa_1_0_0,glue_mrpc_2_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_3epoch_sim_cluster_1 | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0,web_questions_whats_the_answer,web_questions_question_answer,dbpedia_14_pick_one_category_for_the_following_text,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,adversarial_qa_droberta_based_on,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,kilt_tasks_hotpotqa_formulate,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,squad_v1_1_3_0_0 | lora | | phi2_joint_3epoch_sim_cluster_5 | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | phi2_joint_3epoch_sim_cluster_8 | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,ropes_plain_background_situation,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_prompt_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_3epoch_sim_cluster_2 | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,super_glue_record_1_0_2,wiki_hop_original_generate_object,adversarial_qa_droberta_tell_what_it_is,dbpedia_14_given_a_choice_of_categories_,wiki_hop_original_choose_best_object_affirmative_3,quac_1_0_0,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_choose_best_object_affirmative_1,adversarial_qa_dbert_answer_the_following_q,wiki_hop_original_choose_best_object_interrogative_2,adversarial_qa_droberta_question_context_answer,squad_v2_0_3_0_0,wiki_hop_original_generate_subject,wiki_bio_guess_person,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,race_high_Write_a_multi_choice_question_options_given_,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | phi2_joint_3epoch_sim_cluster_7 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_gsm8k,trec_1_0_0,yelp_polarity_reviews_0_2_0,lambada_1_0_0,glue_cola_2_0_0,ag_news_subset_1_0_0,gem_dart_1_1_0,math_dataset_algebra__linear_1d_1_0_0,cnn_dailymail_3_4_0,wiki_hop_original_explain_relation,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,gem_wiki_lingua_english_en_1_1_0,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,gem_web_nlg_en_1_1_0,word_segment,race_high_Write_a_multi_choice_question_for_the_following_article,wmt16_translate_de_en_1_0_0,cot_ecqa,aeslc_1_0_0,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,dream_answer_to_dialogue,para_crawl_enes,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0 | lora | | phi2_joint_3epoch_sim_cluster_6 | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | phi2_joint_3epoch_sim_cluster_4 | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,wmt16_translate_ro_en_1_0_0,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,gem_common_gen_1_1_0,duorc_SelfRC_build_story_around_qa,app_reviews_generate_review,wiki_bio_what_content,wiki_bio_who,gem_e2e_nlg_1_1_0,cot_esnli_ii,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | Last updated on: 2024-01-24T13:39:11.000Z
[ "SCIQ" ]
ostapeno/phi_2-v3-10-flan-clusters_1ep
ostapeno
null
[ "region:us" ]
"2024-01-24T15:20:59Z"
2024-01-24T15:24:21+00:00
0
0
--- {} --- Number of experts present in the library: 5 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_lora_embed_clustersc3_1epoch | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,cos_e_v1_11_question_option_description_id,cot_ecqa_ii,wiqa_effect_with_string_answer,quarel_do_not_use,qasc_qa_with_separated_facts_5,dream_baseline,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,quarel_heres_a_story,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,quarel_testing_students,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,wiqa_effect_with_label_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_lora_embed_clustersc9_1epoch | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,ropes_plain_background_situation,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_prompt_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,quartz_paragraph_question_plain_concat,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_lora_embed_clustersc2_1epoch | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quac_1_0_0,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | phi2_joint_lora_embed_clustersc1_1epoch | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,huggingface_xsum,cot_esnli,cot_gsm8k,gem_dart_1_1_0,gem_common_gen_1_1_0,stream_aqua,cnn_dailymail_3_4_0,gem_wiki_lingua_english_en_1_1_0,gigaword_1_2_0,gem_web_nlg_en_1_1_0,word_segment,app_reviews_generate_review,wmt16_translate_de_en_1_0_0,aeslc_1_0_0,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,dream_answer_to_dialogue,gem_e2e_nlg_1_1_0,para_crawl_enes,wmt14_translate_fr_en_1_0_0,cot_esnli_ii,wiqa_what_is_the_missing_first_step,coqa_1_0_0 | lora | | phi2_joint_lora_embed_clustersc0_1epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,true_case,cot_sensemaking,trec_1_0_0,quartz_answer_question_based_on,cos_e_v1_11_rationale,natural_questions_open_1_0_0,web_questions_whats_the_answer,anli_r3_0_1_0,kilt_tasks_hotpotqa_final_exam,glue_cola_2_0_0,cos_e_v1_11_generate_explanation_given_text,adversarial_qa_droberta_tell_what_it_is,cot_creak,cot_gsm8k_ii,math_dataset_algebra__linear_1d_1_0_0,quartz_use_info_from_question_paragraph,web_questions_question_answer,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,cos_e_v1_11_i_think,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,adversarial_qa_droberta_based_on,web_questions_get_the_answer,adversarial_qa_dbert_answer_the_following_q,kilt_tasks_hotpotqa_complex_question,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,adversarial_qa_droberta_question_context_answer,cos_e_v1_11_explain_why_human,cot_creak_ii,cos_e_v1_11_description_question_option_text,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,stream_qed_ii,wiki_bio_guess_person,cos_e_v1_11_aligned_with_common_sense,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,cot_sensemaking_ii,adversarial_qa_dbert_tell_what_it_is,kilt_tasks_hotpotqa_formulate,social_i_qa_I_was_wondering,cosmos_qa_1_0_0,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is,social_i_qa_Generate_answer,squad_v1_1_3_0_0 | lora | Last updated on: 2024-01-24 15:20:59+00:00 CLusters were obtained using embeddings of experts trained for 5 epochs. THese are the first 5 clusters out of 10.
[ "SCIQ" ]
zhan1993/library-phi_2-v3-10-clusters-loss-one_epoch
zhan1993
null
[ "region:us" ]
"2024-01-25T02:57:28Z"
2024-01-25T02:57:34+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_3epoch_loss_one_epoch_cluster_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1,quarel_testing_students,quartz_use_info_from_paragraph_question,ropes_plain_bottom_hint,ropes_prompt_mix,ropes_read_background_situation,wiki_hop_original_choose_best_object_affirmative_1,quail_no_prompt_text,ropes_given_background_situation,quarel_logic_test,quail_context_question_description_answer_text,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_affirmative_2,wiqa_effect_with_string_answer,quail_context_description_question_answer_text,quartz_read_passage_below_choose,wiki_hop_original_choose_best_object_affirmative_3,quarel_choose_between,sciq_Direct_Question_Closed_Book_,quartz_answer_question_below,ropes_prompt_bottom_no_hint,quartz_given_the_fact_answer_the_q,race_middle_Select_the_best_answer_generate_span_,ropes_background_new_situation_answer,sciq_Direct_Question,social_i_qa_Show_choices_and_generate_answer,quartz_answer_question_based_on,quarel_do_not_use,ropes_prompt_beginning,quarel_heres_a_story,ropes_background_situation_middle,ropes_new_situation_background_answer,ropes_plain_background_situation,quail_description_context_question_answer_text,quartz_paragraph_question_plain_concat,wiqa_which_of_the_following_is_the_supposed_perturbation,quartz_use_info_from_question_paragraph,ropes_plain_no_background,race_high_Select_the_best_answer_generate_span_,ropes_prompt_bottom_hint_beginning | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_7 | phi-2 | sordonia/flan-10k-flat/wiki_bio_comprehension,wiki_bio_what_content,cos_e_v1_11_aligned_with_common_sense,gem_web_nlg_en_1_1_0,cnn_dailymail_3_4_0,race_high_Write_a_multi_choice_question_options_given_,stream_aqua_ii,duorc_SelfRC_build_story_around_qa,cot_esnli_ii,cos_e_v1_11_i_think,stream_qed_ii,cos_e_v1_11_rationale,race_middle_Write_a_multi_choice_question_options_given_,trivia_qa_rc_1_1_0,cot_sensemaking_ii,wiki_bio_key_content,dream_generate_first_utterance,duorc_ParaphraseRC_build_story_around_qa,app_reviews_convert_to_rating,wiqa_what_is_the_final_step_of_the_following_process,cot_strategyqa,wiki_qa_Direct_Answer_to_Question,wiqa_what_might_be_the_first_step_of_the_process,race_middle_Write_a_multi_choice_question_for_the_following_article,wiqa_what_might_be_the_last_step_of_the_process,cos_e_v1_11_generate_explanation_given_text,true_case,cos_e_v1_11_explain_why_human,race_high_Write_a_multi_choice_question_for_the_following_article,huggingface_xsum,cot_creak_ii,wiki_bio_who,dream_answer_to_dialogue,word_segment | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_3 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0,adversarial_qa_dbert_generate_question,quoref_Found_Context_Online,web_questions_get_the_answer,unified_qa_science_inst,quoref_What_Is_The_Answer,adversarial_qa_droberta_generate_question,adversarial_qa_dbidaf_question_context_answer,web_questions_whats_the_answer,adversarial_qa_droberta_question_context_answer,kilt_tasks_hotpotqa_combining_facts,web_questions_potential_correct_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_decide_worth_it,adversarial_qa_droberta_tell_what_it_is,duorc_ParaphraseRC_answer_question,duorc_SelfRC_extract_answer,drop_2_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_dbidaf_based_on,natural_questions_open_1_0_0,lambada_1_0_0,quoref_Read_And_Extract_,duorc_SelfRC_question_answering,squad_v1_1_3_0_0,coqa_1_0_0,web_questions_question_answer,wiki_qa_Topic_Prediction_Question_Only,quoref_Find_Answer,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,cos_e_v1_11_question_description_option_text,duorc_SelfRC_movie_director,quoref_Given_Context_Answer_Question,adversarial_qa_dbidaf_tell_what_it_is,web_questions_short_general_knowledge_q,duorc_SelfRC_answer_question,quoref_Answer_Question_Given_Context,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_text,adversarial_qa_dbert_based_on,duorc_ParaphraseRC_movie_director,quoref_Guess_Answer,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,kilt_tasks_hotpotqa_final_exam,quoref_Answer_Friend_Question,adversarial_qa_dbert_question_context_answer,kilt_tasks_hotpotqa_complex_question,wiki_qa_Topic_Prediction_Answer_Only,duorc_ParaphraseRC_decide_worth_it,quoref_Answer_Test,kilt_tasks_hotpotqa_formulate,gigaword_1_2_0,duorc_ParaphraseRC_question_answering,adversarial_qa_droberta_based_on | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_5 | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test,quail_no_prompt_id,quail_context_question_answer_description_id,race_high_Select_the_best_answer,race_high_Select_the_best_answer_no_instructions_,wiqa_effect_with_label_answer,race_middle_Select_the_best_answer,cos_e_v1_11_question_description_option_id,quail_context_description_question_answer_id,quail_description_context_question_answer_id,wiki_qa_Decide_good_answer,quail_context_question_description_answer_id,race_middle_Select_the_best_answer_no_instructions_,cos_e_v1_11_description_question_option_id,race_middle_Taking_a_test,cos_e_v1_11_question_option_description_id,social_i_qa_Show_choices_and_generate_index | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_9 | phi-2 | sordonia/flan-10k-flat/stream_aqua,cot_esnli,fix_punct,cot_qasc,cot_gsm8k,multi_news_1_0_0,cot_creak,stream_qed,cot_sensemaking,cot_ecqa | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_1 | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice,qasc_qa_with_separated_facts_1,qasc_qa_with_separated_facts_2,quartz_having_read_above_passage,dream_baseline,sciq_Multiple_Choice_Closed_Book_,qasc_qa_with_separated_facts_4,quail_context_question_answer_description_text,sciq_Multiple_Choice_Question_First,dream_read_the_following_conversation_and_answer_the_question,qasc_qa_with_combined_facts_1 | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_6 | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_,race_high_Read_the_article_and_answer_the_question_no_option_,yelp_polarity_reviews_0_2_0,qasc_is_correct_1,social_i_qa_Generate_answer,cosmos_qa_1_0_0,qasc_qa_with_separated_facts_5,glue_stsb_2_0_0,social_i_qa_I_was_wondering,qasc_qa_with_separated_facts_3,quail_description_context_question_text,qasc_is_correct_2,quail_context_question_description_text,social_i_qa_Check_if_a_random_answer_is_valid_or_not,race_middle_Read_the_article_and_answer_the_question_no_option_,quail_context_description_question_text,wiki_qa_automatic_system | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_8 | phi-2 | sordonia/flan-10k-flat/trec_1_0_0,glue_qqp_2_0_0,paws_wiki_1_1_0,anli_r3_0_1_0,anli_r2_0_1_0,anli_r1_0_1_0,ag_news_subset_1_0_0 | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_10 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,imdb_reviews_plain_text_1_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,snli_1_1_0,super_glue_multirc_1_0_2,super_glue_wic_1_0_2,definite_pronoun_resolution_1_1_0,super_glue_rte_1_0_2,glue_mrpc_2_0_0,glue_wnli_2_0_0,super_glue_wsc_fixed_1_0_2,glue_cola_2_0_0 | lora | | phi2_joint_3epoch_loss_one_epoch_cluster_4 | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise,duorc_SelfRC_generate_question_by_answer,cot_gsm8k_ii,gem_wiki_lingua_english_en_1_1_0,wiki_hop_original_generate_object,wiqa_what_is_the_missing_first_step,wiki_qa_found_on_google,wmt16_translate_de_en_1_0_0,quoref_Guess_Title_For_Context,gem_common_gen_1_1_0,super_glue_cb_1_0_2,para_crawl_enes,app_reviews_generate_review,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,super_glue_copa_1_0_2,math_dataset_algebra__linear_1d_1_0_0,app_reviews_convert_to_star_rating,wiki_bio_guess_person,wiqa_does_the_supposed_perturbation_have_an_effect,wiki_qa_Jeopardy_style,wmt16_translate_tr_en_1_0_0,duorc_ParaphraseRC_title_generation,race_middle_Is_this_the_right_answer,app_reviews_categorize_rating_using_review,gem_dart_1_1_0,duorc_ParaphraseRC_generate_question_by_answer,wiki_hop_original_generate_subject,aeslc_1_0_0,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,wmt16_translate_ro_en_1_0_0,race_high_Is_this_the_right_answer,duorc_SelfRC_title_generation,duorc_SelfRC_generate_question,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_hop_original_generate_subject_and_object,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_explain_relation,super_glue_record_1_0_2,cot_ecqa_ii,wmt14_translate_fr_en_1_0_0,dbpedia_14_given_a_choice_of_categories_,wiki_qa_Generate_Question_from_Topic,dbpedia_14_pick_one_category_for_the_following_text,gem_e2e_nlg_1_1_0,social_i_qa_Generate_the_question_from_the_answer,cot_strategyqa_ii,wmt16_translate_fi_en_1_0_0,duorc_ParaphraseRC_generate_question,dream_generate_last_utterance,quac_1_0_0 | lora | Last updated on: 2024-01-25T02:57:28.000Z
[ "SCIQ" ]
zhan1993/gptneo_1B_flan_random-epoch_0
zhan1993
null
[ "region:us" ]
"2024-01-25T20:55:19Z"
2024-01-26T15:00:44+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5,wiki_qa_automatic_system,stream_aqua_ii,dbpedia_14_pick_one_category_for_the_following_text,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Generate_Question_from_Topic,cot_gsm8k_ii,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_interrogative_2,quail_context_question_answer_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,qasc_qa_with_combined_facts_1,adversarial_qa_dbert_answer_the_following_q,social_i_qa_I_was_wondering,stream_aqua,word_segment,ropes_plain_no_background,super_glue_multirc_1_0_2,wiki_hop_original_choose_best_object_affirmative_3,app_reviews_convert_to_rating,anli_r3_0_1_0,app_reviews_convert_to_star_rating,quartz_paragraph_question_plain_concat,kilt_tasks_hotpotqa_complex_question,quartz_use_info_from_paragraph_question | lora | | cluster_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0,para_crawl_enes,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_id,glue_stsb_2_0_0,quail_context_description_question_answer_text,dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_description_option_text,cos_e_v1_11_description_question_option_text,race_high_Write_a_multi_choice_question_options_given_,ropes_prompt_bottom_no_hint,quarel_testing_students,wmt16_translate_tr_en_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_final_step_of_the_following_process,quoref_Given_Context_Answer_Question,wiki_hop_original_generate_object,quartz_use_info_from_question_paragraph,duorc_SelfRC_build_story_around_qa,drop_2_0_0,wiqa_effect_with_string_answer,race_high_Taking_a_test,wiki_hop_original_generate_subject_and_object,glue_qqp_2_0_0,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | cluster_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0,duorc_SelfRC_question_answering,cot_sensemaking_ii,huggingface_xsum,duorc_ParaphraseRC_build_story_around_qa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,quartz_answer_question_based_on,ropes_plain_background_situation,race_middle_Write_a_multi_choice_question_for_the_following_article,quail_description_context_question_text,web_questions_whats_the_answer,cot_ecqa,true_case,adversarial_qa_dbert_question_context_answer,duorc_SelfRC_title_generation,quail_context_question_description_answer_id,quarel_do_not_use,adversarial_qa_dbidaf_answer_the_following_q,duorc_ParaphraseRC_decide_worth_it,race_middle_Is_this_the_right_answer,wmt16_translate_de_en_1_0_0,wiki_qa_Is_This_True_,race_middle_Select_the_best_answer,aeslc_1_0_0,duorc_ParaphraseRC_question_answering,wiki_hop_original_choose_best_object_affirmative_1 | lora | | cluster_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_explain_why_human,cot_sensemaking,quoref_Find_Answer,quail_context_description_question_text,social_i_qa_Generate_the_question_from_the_answer,quartz_having_read_above_passage,stream_qed_ii,wiqa_what_might_be_the_first_step_of_the_process,wiki_bio_who,duorc_SelfRC_generate_question_by_answer,race_high_Select_the_best_answer_generate_span_,lambada_1_0_0,coqa_1_0_0,race_high_Select_the_best_answer,adversarial_qa_droberta_question_context_answer,quail_context_question_answer_description_id,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_tell_what_it_is,cnn_dailymail_3_4_0,quail_no_prompt_id,ag_news_subset_1_0_0,trivia_qa_rc_1_1_0,ropes_prompt_bottom_hint_beginning,super_glue_wic_1_0_2 | lora | | cluster_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2,duorc_ParaphraseRC_movie_director,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,quartz_read_passage_below_choose,social_i_qa_Check_if_a_random_answer_is_valid_or_not,cot_creak,ropes_plain_bottom_hint,super_glue_copa_1_0_2,natural_questions_open_1_0_0,trec_1_0_0,gem_web_nlg_en_1_1_0,wiki_bio_key_content,wmt16_translate_fi_en_1_0_0,quoref_Answer_Question_Given_Context,duorc_ParaphraseRC_generate_question,math_dataset_algebra__linear_1d_1_0_0,duorc_ParaphraseRC_title_generation,quail_context_description_question_answer_id,wiki_bio_what_content,adversarial_qa_dbert_tell_what_it_is,sciq_Multiple_Choice_Closed_Book_,duorc_ParaphraseRC_extract_answer,dream_baseline,gem_wiki_lingua_english_en_1_1_0 | lora | | cluster_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id,social_i_qa_Show_choices_and_generate_index,web_questions_potential_correct_answer,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cosmos_qa_1_0_0,sciq_Direct_Question,super_glue_wsc_fixed_1_0_2,race_middle_Taking_a_test,wmt14_translate_fr_en_1_0_0,duorc_SelfRC_extract_answer,wiki_hop_original_generate_subject,duorc_SelfRC_answer_question,qasc_is_correct_1,cos_e_v1_11_i_think,wiki_qa_exercise,race_middle_Write_a_multi_choice_question_options_given_,quoref_Read_And_Extract_,web_questions_short_general_knowledge_q,web_questions_question_answer,quarel_logic_test,app_reviews_categorize_rating_using_review,cot_strategyqa_ii,glue_mnli_2_0_0,quoref_Answer_Test,super_glue_rte_1_0_2 | lora | | cluster_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed,cot_esnli_ii,quarel_heres_a_story,quoref_Guess_Title_For_Context,qasc_is_correct_2,wiqa_effect_with_label_answer,dream_generate_last_utterance,adversarial_qa_dbert_based_on,dream_answer_to_dialogue,sciq_Multiple_Choice_Question_First,quail_context_question_description_answer_text,wiki_qa_Direct_Answer_to_Question,ropes_background_new_situation_answer,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,quoref_What_Is_The_Answer,dbpedia_14_given_a_choice_of_categories_,qasc_qa_with_separated_facts_2,glue_mrpc_2_0_0,gem_e2e_nlg_1_1_0,anli_r1_0_1_0,race_middle_Read_the_article_and_answer_the_question_no_option_,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,adversarial_qa_dbidaf_based_on | lora | | cluster_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii,quail_description_context_question_answer_id,kilt_tasks_hotpotqa_final_exam,cot_gsm8k,cos_e_v1_11_aligned_with_common_sense,squad_v2_0_3_0_0,duorc_SelfRC_movie_director,anli_r2_0_1_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,duorc_ParaphraseRC_answer_question,ropes_background_situation_middle,dream_generate_first_utterance,quail_context_question_description_text,adversarial_qa_droberta_based_on,glue_wnli_2_0_0,super_glue_record_1_0_2,web_questions_get_the_answer,ropes_prompt_mix,app_reviews_generate_review,cos_e_v1_11_rationale,adversarial_qa_droberta_generate_question,yelp_polarity_reviews_0_2_0,ropes_given_background_situation,qasc_qa_with_separated_facts_3,quoref_Guess_Answer | lora | | cluster_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article,cot_qasc,definite_pronoun_resolution_1_1_0,wiki_qa_found_on_google,wiki_bio_comprehension,wiki_qa_Topic_Prediction_Question_Only,wiki_bio_guess_person,fix_punct,race_middle_Select_the_best_answer_no_instructions_,quac_1_0_0,wiqa_does_the_supposed_perturbation_have_an_effect,quartz_given_the_fact_answer_the_q,ropes_new_situation_background_answer,social_i_qa_Generate_answer,gigaword_1_2_0,duorc_SelfRC_decide_worth_it,kilt_tasks_hotpotqa_straighforward_qa,quail_no_prompt_text,cot_esnli,quoref_Answer_Friend_Question,race_middle_Select_the_best_answer_generate_span_,unified_qa_science_inst,sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_strategyqa | lora | | cluster_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0,quarel_choose_between,wiki_hop_original_explain_relation,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,race_high_Select_the_best_answer_no_instructions_,kilt_tasks_hotpotqa_combining_facts,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_generate_question,glue_cola_2_0_0,imdb_reviews_plain_text_1_0_0,squad_v1_1_3_0_0,race_high_Is_this_the_right_answer,qasc_qa_with_separated_facts_1,glue_sst2_2_0_0,wiqa_what_is_the_missing_first_step,duorc_ParaphraseRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,wiki_qa_Jeopardy_style,quartz_answer_question_below,ropes_prompt_beginning,ropes_read_background_situation,wiki_qa_Decide_good_answer,super_glue_cb_1_0_2,qasc_qa_with_separated_facts_4,cot_ecqa_ii | lora | Last updated on: 2024-01-26T14:57:09.000Z
[ "SCIQ" ]
ostapeno/library-gptneo_1B_flan_5ep_underparam
ostapeno
null
[ "region:us" ]
"2024-01-25T23:24:02Z"
2024-02-08T18:50:18+00:00
0
0
--- {} --- Number of experts present in the library: 256 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | aeslc_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/aeslc_1_0_0 | lora | | race_middle_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Is_this_the_right_answer | lora | | true_case | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/true_case | lora | | cot_strategyqa_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_strategyqa_ii | lora | | quarel_do_not_use | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quarel_do_not_use | lora | | word_segment | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Decide_good_answer | lora | | stream_aqua_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/stream_aqua_ii | lora | | wiki_hop_original_choose_best_object_affirmative_3 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | duorc_ParaphraseRC_movie_director | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_movie_director | lora | | math_dataset_algebra__linear_1d_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cot_gsm8k_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_gsm8k_ii | lora | | social_i_qa_Show_choices_and_generate_index | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dream_generate_last_utterance | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dream_generate_last_utterance | lora | | quail_context_question_description_answer_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_question_description_answer_id | lora | | cot_creak_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_creak_ii | lora | | ropes_background_new_situation_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | cot_esnli | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_esnli | lora | | anli_r3_0_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/anli_r3_0_1_0 | lora | | adversarial_qa_droberta_question_context_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | wiki_bio_who | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_bio_who | lora | | cos_e_v1_11_i_think | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_i_think | lora | | gem_wiki_lingua_english_en_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_prompt_bottom_no_hint | lora | | glue_qqp_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_description_question_option_text | lora | | wiki_qa_Topic_Prediction_Answer_Only | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | cot_creak | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_creak | lora | | trivia_qa_rc_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/trivia_qa_rc_1_1_0 | lora | | duorc_SelfRC_title_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_title_generation | lora | | wiqa_what_is_the_final_step_of_the_following_process | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | glue_mnli_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_mnli_2_0_0 | lora | | quail_context_question_description_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Guess_Answer | lora | | adversarial_qa_dbidaf_generate_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | adversarial_qa_droberta_answer_the_following_q | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | cos_e_v1_11_question_description_option_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_question_description_option_id | lora | | wiki_hop_original_generate_subject_and_object | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | para_crawl_enes | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/para_crawl_enes | lora | | ropes_background_situation_middle | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_background_situation_middle | lora | | adversarial_qa_dbert_generate_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbert_generate_question | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | ropes_prompt_beginning | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_prompt_beginning | lora | | cot_strategyqa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_strategyqa | lora | | duorc_ParaphraseRC_question_answering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_question_answering | lora | | adversarial_qa_dbert_based_on | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_hop_original_generate_subject | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_generate_subject | lora | | race_high_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | cot_gsm8k | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_gsm8k | lora | | glue_mrpc_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_mrpc_2_0_0 | lora | | cos_e_v1_11_explain_why_human | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_explain_why_human | lora | | race_high_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | app_reviews_categorize_rating_using_review | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/app_reviews_categorize_rating_using_review | lora | | adversarial_qa_dbert_answer_the_following_q | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | quail_context_question_answer_description_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_question_answer_description_text | lora | | qasc_qa_with_combined_facts_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_combined_facts_1 | lora | | quoref_What_Is_The_Answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_description_question_answer_text | lora | | glue_stsb_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_stsb_2_0_0 | lora | | qasc_qa_with_separated_facts_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_separated_facts_1 | lora | | web_questions_short_general_knowledge_q | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quarel_testing_students | lora | | cnn_dailymail_3_4_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cnn_dailymail_3_4_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Multiple_Choice | lora | | kilt_tasks_hotpotqa_complex_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | huggingface_xsum | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/huggingface_xsum | lora | | cot_sensemaking_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_sensemaking_ii | lora | | glue_sst2_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_sst2_2_0_0 | lora | | dbpedia_14_pick_one_category_for_the_following_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | qasc_is_correct_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_is_correct_2 | lora | | wiki_qa_Generate_Question_from_Topic | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_combining_facts | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quartz_given_the_fact_answer_the_q | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_given_the_fact_answer_the_q | lora | | wmt16_translate_ro_en_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | cot_sensemaking | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_sensemaking | lora | | wiki_bio_what_content | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/sciq_Direct_Question | lora | | quail_context_question_description_answer_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_prompt_mix | lora | | quartz_answer_question_based_on | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_answer_question_based_on | lora | | fix_punct | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/fix_punct | lora | | qasc_is_correct_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_is_correct_1 | lora | | cos_e_v1_11_question_description_option_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_question_description_option_text | lora | | wmt16_translate_fi_en_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_multirc_1_0_2 | lora | | wiki_qa_Is_This_True_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/multi_news_1_0_0 | lora | | web_questions_question_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/web_questions_question_answer | lora | | quartz_use_info_from_paragraph_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_choose_between | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quarel_choose_between | lora | | adversarial_qa_droberta_tell_what_it_is | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | cot_ecqa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_ecqa | lora | | web_questions_get_the_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_movie_director | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_movie_director | lora | | dream_answer_to_dialogue | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dream_answer_to_dialogue | lora | | ropes_plain_background_situation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_plain_background_situation | lora | | cos_e_v1_11_rationale | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_rationale | lora | | duorc_ParaphraseRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_exercise | lora | | kilt_tasks_hotpotqa_straighforward_qa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | wiki_hop_original_explain_relation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | natural_questions_open_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/natural_questions_open_1_0_0 | lora | | anli_r1_0_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/anli_r1_0_1_0 | lora | | quoref_Given_Context_Answer_Question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Given_Context_Answer_Question | lora | | duorc_SelfRC_answer_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_answer_question | lora | | wmt16_translate_de_en_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_generate_object | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | duorc_ParaphraseRC_extract_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | unified_qa_science_inst | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_description_context_question_text | lora | | yelp_polarity_reviews_0_2_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | quarel_logic_test | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Jeopardy_style | lora | | wiki_qa_automatic_system | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | duorc_SelfRC_build_story_around_qa | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Answer_Friend_Question | lora | | quartz_having_read_above_passage | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_having_read_above_passage | lora | | glue_cola_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_cola_2_0_0 | lora | | wiqa_effect_with_string_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_effect_with_string_answer | lora | | duorc_ParaphraseRC_answer_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_answer_question | lora | | ag_news_subset_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ag_news_subset_1_0_0 | lora | | dbpedia_14_given_a_choice_of_categories_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | dream_baseline | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dream_baseline | lora | | qasc_qa_with_separated_facts_4 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_read_passage_below_choose | lora | | race_middle_Write_a_multi_choice_question_options_given_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_plain_no_background | lora | | qasc_qa_with_separated_facts_5 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Read_And_Extract_ | lora | | duorc_ParaphraseRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | duorc_SelfRC_question_answering | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_question_answering | lora | | definite_pronoun_resolution_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | super_glue_rte_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | stream_qed | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/stream_qed | lora | | app_reviews_convert_to_rating | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/app_reviews_convert_to_rating | lora | | wiqa_what_might_be_the_first_step_of_the_process | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | social_i_qa_Show_choices_and_generate_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | cos_e_v1_11_generate_explanation_given_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | quartz_use_info_from_question_paragraph | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_use_info_from_question_paragraph | lora | | anli_r2_0_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/anli_r2_0_1_0 | lora | | duorc_ParaphraseRC_generate_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quarel_heres_a_story | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quarel_heres_a_story | lora | | cos_e_v1_11_aligned_with_common_sense | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | duorc_SelfRC_extract_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_extract_answer | lora | | race_middle_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | web_questions_potential_correct_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_separated_facts_3 | lora | | adversarial_qa_dbidaf_tell_what_it_is | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_qasc | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_qasc | lora | | adversarial_qa_dbert_tell_what_it_is | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | paws_wiki_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/paws_wiki_1_1_0 | lora | | quail_context_description_question_answer_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | kilt_tasks_hotpotqa_final_exam | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | trec_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_effect_with_label_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | snli_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/snli_1_1_0 | lora | | cot_ecqa_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_ecqa_ii | lora | | quail_context_question_answer_description_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_question_answer_description_id | lora | | gigaword_1_2_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gigaword_1_2_0 | lora | | cos_e_v1_11_question_option_description_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_question_option_description_text | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | glue_qnli_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_qnli_2_0_0 | lora | | duorc_SelfRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | kilt_tasks_hotpotqa_formulate | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | quartz_paragraph_question_plain_concat | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_paragraph_question_plain_concat | lora | | adversarial_qa_dbert_question_context_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | ropes_prompt_bottom_hint_beginning | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | adversarial_qa_droberta_based_on | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_droberta_based_on | lora | | dream_generate_first_utterance | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dream_generate_first_utterance | lora | | duorc_SelfRC_decide_worth_it | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_decide_worth_it | lora | | quail_context_description_question_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_context_description_question_text | lora | | race_high_Is_this_the_right_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/drop_2_0_0 | lora | | ropes_read_background_situation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_read_background_situation | lora | | glue_wnli_2_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/glue_wnli_2_0_0 | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | wiki_qa_Direct_Answer_to_Question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | web_questions_whats_the_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_found_on_google | lora | | quail_no_prompt_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_no_prompt_text | lora | | duorc_ParaphraseRC_generate_question_by_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | cos_e_v1_11_question_option_description_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_question_option_description_id | lora | | super_glue_copa_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_copa_1_0_2 | lora | | app_reviews_generate_review | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/app_reviews_generate_review | lora | | wiki_hop_original_choose_best_object_affirmative_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | app_reviews_convert_to_star_rating | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/app_reviews_convert_to_star_rating | lora | | gem_web_nlg_en_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gem_web_nlg_en_1_1_0 | lora | | quoref_Context_Contains_Answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Context_Contains_Answer | lora | | gem_e2e_nlg_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gem_e2e_nlg_1_1_0 | lora | | quoref_Answer_Question_Given_Context | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quartz_answer_question_below | lora | | duorc_SelfRC_generate_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_SelfRC_generate_question | lora | | race_high_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Select_the_best_answer | lora | | stream_qed_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/stream_qed_ii | lora | | cos_e_v1_11_description_question_option_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_bio_comprehension | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_bio_comprehension | lora | | duorc_ParaphraseRC_title_generation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_no_prompt_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_no_prompt_id | lora | | adversarial_qa_dbidaf_answer_the_following_q | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | quoref_Found_Context_Online | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Select_the_best_answer | lora | | race_middle_Taking_a_test | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Answer_Test | lora | | gem_common_gen_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gem_common_gen_1_1_0 | lora | | race_high_Taking_a_test | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_description_context_question_answer_id | lora | | gem_dart_1_1_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/gem_dart_1_1_0 | lora | | wiki_bio_guess_person | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quac_1_0_0 | lora | | cosmos_qa_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cosmos_qa_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | lambada_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/lambada_1_0_0 | lora | | ropes_given_background_situation | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | adversarial_qa_droberta_generate_question | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_droberta_generate_question | lora | | adversarial_qa_dbidaf_question_context_answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | adversarial_qa_dbidaf_based_on | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiki_hop_original_choose_best_object_affirmative_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | cot_esnli_ii | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/cot_esnli_ii | lora | | quail_description_context_question_answer_text | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/super_glue_record_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | EleutherAI/gpt-neo-1.3B | ostapeno/adauni-v3-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | Last updated on: 2024-01-26 17:15:01+00:00
[ "SCIQ" ]
zhan1993/gptneo_1B_flan_random-epoch_1
zhan1993
null
[ "region:us" ]
"2024-01-26T00:53:18Z"
2024-01-26T15:02:22+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5,wiki_qa_automatic_system,stream_aqua_ii,dbpedia_14_pick_one_category_for_the_following_text,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Generate_Question_from_Topic,cot_gsm8k_ii,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_interrogative_2,quail_context_question_answer_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,qasc_qa_with_combined_facts_1,adversarial_qa_dbert_answer_the_following_q,social_i_qa_I_was_wondering,stream_aqua,word_segment,ropes_plain_no_background,super_glue_multirc_1_0_2,wiki_hop_original_choose_best_object_affirmative_3,app_reviews_convert_to_rating,anli_r3_0_1_0,app_reviews_convert_to_star_rating,quartz_paragraph_question_plain_concat,kilt_tasks_hotpotqa_complex_question,quartz_use_info_from_paragraph_question | lora | | cluster_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0,para_crawl_enes,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_id,glue_stsb_2_0_0,quail_context_description_question_answer_text,dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_description_option_text,cos_e_v1_11_description_question_option_text,race_high_Write_a_multi_choice_question_options_given_,ropes_prompt_bottom_no_hint,quarel_testing_students,wmt16_translate_tr_en_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_final_step_of_the_following_process,quoref_Given_Context_Answer_Question,wiki_hop_original_generate_object,quartz_use_info_from_question_paragraph,duorc_SelfRC_build_story_around_qa,drop_2_0_0,wiqa_effect_with_string_answer,race_high_Taking_a_test,wiki_hop_original_generate_subject_and_object,glue_qqp_2_0_0,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | cluster_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0,duorc_SelfRC_question_answering,cot_sensemaking_ii,huggingface_xsum,duorc_ParaphraseRC_build_story_around_qa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,quartz_answer_question_based_on,ropes_plain_background_situation,race_middle_Write_a_multi_choice_question_for_the_following_article,quail_description_context_question_text,web_questions_whats_the_answer,cot_ecqa,true_case,adversarial_qa_dbert_question_context_answer,duorc_SelfRC_title_generation,quail_context_question_description_answer_id,quarel_do_not_use,adversarial_qa_dbidaf_answer_the_following_q,duorc_ParaphraseRC_decide_worth_it,race_middle_Is_this_the_right_answer,wmt16_translate_de_en_1_0_0,wiki_qa_Is_This_True_,race_middle_Select_the_best_answer,aeslc_1_0_0,duorc_ParaphraseRC_question_answering,wiki_hop_original_choose_best_object_affirmative_1 | lora | | cluster_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_explain_why_human,cot_sensemaking,quoref_Find_Answer,quail_context_description_question_text,social_i_qa_Generate_the_question_from_the_answer,quartz_having_read_above_passage,stream_qed_ii,wiqa_what_might_be_the_first_step_of_the_process,wiki_bio_who,duorc_SelfRC_generate_question_by_answer,race_high_Select_the_best_answer_generate_span_,lambada_1_0_0,coqa_1_0_0,race_high_Select_the_best_answer,adversarial_qa_droberta_question_context_answer,quail_context_question_answer_description_id,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_tell_what_it_is,cnn_dailymail_3_4_0,quail_no_prompt_id,ag_news_subset_1_0_0,trivia_qa_rc_1_1_0,ropes_prompt_bottom_hint_beginning,super_glue_wic_1_0_2 | lora | | cluster_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2,duorc_ParaphraseRC_movie_director,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,quartz_read_passage_below_choose,social_i_qa_Check_if_a_random_answer_is_valid_or_not,cot_creak,ropes_plain_bottom_hint,super_glue_copa_1_0_2,natural_questions_open_1_0_0,trec_1_0_0,gem_web_nlg_en_1_1_0,wiki_bio_key_content,wmt16_translate_fi_en_1_0_0,quoref_Answer_Question_Given_Context,duorc_ParaphraseRC_generate_question,math_dataset_algebra__linear_1d_1_0_0,duorc_ParaphraseRC_title_generation,quail_context_description_question_answer_id,wiki_bio_what_content,adversarial_qa_dbert_tell_what_it_is,sciq_Multiple_Choice_Closed_Book_,duorc_ParaphraseRC_extract_answer,dream_baseline,gem_wiki_lingua_english_en_1_1_0 | lora | | cluster_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0,quarel_choose_between,wiki_hop_original_explain_relation,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,race_high_Select_the_best_answer_no_instructions_,kilt_tasks_hotpotqa_combining_facts,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_generate_question,glue_cola_2_0_0,imdb_reviews_plain_text_1_0_0,squad_v1_1_3_0_0,race_high_Is_this_the_right_answer,qasc_qa_with_separated_facts_1,glue_sst2_2_0_0,wiqa_what_is_the_missing_first_step,duorc_ParaphraseRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,wiki_qa_Jeopardy_style,quartz_answer_question_below,ropes_prompt_beginning,ropes_read_background_situation,wiki_qa_Decide_good_answer,super_glue_cb_1_0_2,qasc_qa_with_separated_facts_4,cot_ecqa_ii | lora | | cluster_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed,cot_esnli_ii,quarel_heres_a_story,quoref_Guess_Title_For_Context,qasc_is_correct_2,wiqa_effect_with_label_answer,dream_generate_last_utterance,adversarial_qa_dbert_based_on,dream_answer_to_dialogue,sciq_Multiple_Choice_Question_First,quail_context_question_description_answer_text,wiki_qa_Direct_Answer_to_Question,ropes_background_new_situation_answer,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,quoref_What_Is_The_Answer,dbpedia_14_given_a_choice_of_categories_,qasc_qa_with_separated_facts_2,glue_mrpc_2_0_0,gem_e2e_nlg_1_1_0,anli_r1_0_1_0,race_middle_Read_the_article_and_answer_the_question_no_option_,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,adversarial_qa_dbidaf_based_on | lora | | cluster_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii,quail_description_context_question_answer_id,kilt_tasks_hotpotqa_final_exam,cot_gsm8k,cos_e_v1_11_aligned_with_common_sense,squad_v2_0_3_0_0,duorc_SelfRC_movie_director,anli_r2_0_1_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,duorc_ParaphraseRC_answer_question,ropes_background_situation_middle,dream_generate_first_utterance,quail_context_question_description_text,adversarial_qa_droberta_based_on,glue_wnli_2_0_0,super_glue_record_1_0_2,web_questions_get_the_answer,ropes_prompt_mix,app_reviews_generate_review,cos_e_v1_11_rationale,adversarial_qa_droberta_generate_question,yelp_polarity_reviews_0_2_0,ropes_given_background_situation,qasc_qa_with_separated_facts_3,quoref_Guess_Answer | lora | | cluster_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article,cot_qasc,definite_pronoun_resolution_1_1_0,wiki_qa_found_on_google,wiki_bio_comprehension,wiki_qa_Topic_Prediction_Question_Only,wiki_bio_guess_person,fix_punct,race_middle_Select_the_best_answer_no_instructions_,quac_1_0_0,wiqa_does_the_supposed_perturbation_have_an_effect,quartz_given_the_fact_answer_the_q,ropes_new_situation_background_answer,social_i_qa_Generate_answer,gigaword_1_2_0,duorc_SelfRC_decide_worth_it,kilt_tasks_hotpotqa_straighforward_qa,quail_no_prompt_text,cot_esnli,quoref_Answer_Friend_Question,race_middle_Select_the_best_answer_generate_span_,unified_qa_science_inst,sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_strategyqa | lora | | cluster_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id,social_i_qa_Show_choices_and_generate_index,web_questions_potential_correct_answer,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cosmos_qa_1_0_0,sciq_Direct_Question,super_glue_wsc_fixed_1_0_2,race_middle_Taking_a_test,wmt14_translate_fr_en_1_0_0,duorc_SelfRC_extract_answer,wiki_hop_original_generate_subject,duorc_SelfRC_answer_question,qasc_is_correct_1,cos_e_v1_11_i_think,wiki_qa_exercise,race_middle_Write_a_multi_choice_question_options_given_,quoref_Read_And_Extract_,web_questions_short_general_knowledge_q,web_questions_question_answer,quarel_logic_test,app_reviews_categorize_rating_using_review,cot_strategyqa_ii,glue_mnli_2_0_0,quoref_Answer_Test,super_glue_rte_1_0_2 | lora | Last updated on: 2024-01-26T14:59:27.000Z
[ "SCIQ" ]
ntc-ai/SDXL-LoRA-slider.wearing-a-suit-and-tie
ntc-ai
text-to-image
[ "diffusers", "text-to-image", "stable-diffusion-xl", "lora", "template:sd-lora", "template:sdxl-lora", "sdxl-sliders", "ntcai.xyz-sliders", "concept", "en", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:mit", "region:us" ]
"2024-01-26T01:27:49Z"
2024-01-26T01:27:57+00:00
0
0
--- base_model: stabilityai/stable-diffusion-xl-base-1.0 language: - en license: mit tags: - text-to-image - stable-diffusion-xl - lora - template:sd-lora - template:sdxl-lora - sdxl-sliders - ntcai.xyz-sliders - concept - diffusers thumbnail: images/evaluate/wearing a suit and tie.../wearing a suit and tie_17_3.0.png widget: - text: wearing a suit and tie output: url: images/wearing a suit and tie_17_3.0.png - text: wearing a suit and tie output: url: images/wearing a suit and tie_19_3.0.png - text: wearing a suit and tie output: url: images/wearing a suit and tie_20_3.0.png - text: wearing a suit and tie output: url: images/wearing a suit and tie_21_3.0.png - text: wearing a suit and tie output: url: images/wearing a suit and tie_22_3.0.png inference: false instance_prompt: wearing a suit and tie --- # ntcai.xyz slider - wearing a suit and tie (SDXL LoRA) | Strength: -3 | Strength: 0 | Strength: 3 | | --- | --- | --- | | <img src="images/wearing a suit and tie_17_-3.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_17_0.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_17_3.0.png" width=256 height=256 /> | | <img src="images/wearing a suit and tie_19_-3.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_19_0.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_19_3.0.png" width=256 height=256 /> | | <img src="images/wearing a suit and tie_20_-3.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_20_0.0.png" width=256 height=256 /> | <img src="images/wearing a suit and tie_20_3.0.png" width=256 height=256 /> | ## Download Weights for this model are available in Safetensors format. ## Trigger words You can apply this LoRA with trigger words for additional effect: ``` wearing a suit and tie ``` ## Use in diffusers ```python from diffusers import StableDiffusionXLPipeline from diffusers import EulerAncestralDiscreteScheduler import torch pipe = StableDiffusionXLPipeline.from_single_file("https://huggingface.co/martyn/sdxl-turbo-mario-merge-top-rated/blob/main/topRatedTurboxlLCM_v10.safetensors") pipe.to("cuda") pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) # Load the LoRA pipe.load_lora_weights('ntc-ai/SDXL-LoRA-slider.wearing-a-suit-and-tie', weight_name='wearing a suit and tie.safetensors', adapter_name="wearing a suit and tie") # Activate the LoRA pipe.set_adapters(["wearing a suit and tie"], adapter_weights=[2.0]) prompt = "medieval rich kingpin sitting in a tavern, wearing a suit and tie" negative_prompt = "nsfw" width = 512 height = 512 num_inference_steps = 10 guidance_scale = 2 image = pipe(prompt, negative_prompt=negative_prompt, width=width, height=height, guidance_scale=guidance_scale, num_inference_steps=num_inference_steps).images[0] image.save('result.png') ``` ## Support the Patreon If you like this model please consider [joining our Patreon](https://www.patreon.com/NTCAI). By joining our Patreon, you'll gain access to an ever-growing library of over 1140+ unique and diverse LoRAs, covering a wide range of styles and genres. You'll also receive early access to new models and updates, exclusive behind-the-scenes content, and the powerful LoRA slider creator, allowing you to craft your own custom LoRAs and experiment with endless possibilities. Your support on Patreon will allow us to continue developing and refining new models. ## Other resources - [CivitAI](https://civitai.com/user/ntc) - Follow ntc on Civit for even more LoRAs - [ntcai.xyz](https://ntcai.xyz) - See ntcai.xyz to find more articles and LoRAs
[ "CRAFT" ]
zhan1993/gptneo_1B_flan_random-epoch_2
zhan1993
null
[ "region:us" ]
"2024-01-26T03:22:38Z"
2024-01-26T17:08:25+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5,wiki_qa_automatic_system,stream_aqua_ii,dbpedia_14_pick_one_category_for_the_following_text,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Generate_Question_from_Topic,cot_gsm8k_ii,cos_e_v1_11_question_option_description_id,cos_e_v1_11_question_option_description_text,wiki_hop_original_choose_best_object_interrogative_2,quail_context_question_answer_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,qasc_qa_with_combined_facts_1,adversarial_qa_dbert_answer_the_following_q,social_i_qa_I_was_wondering,stream_aqua,word_segment,ropes_plain_no_background,super_glue_multirc_1_0_2,wiki_hop_original_choose_best_object_affirmative_3,app_reviews_convert_to_rating,anli_r3_0_1_0,app_reviews_convert_to_star_rating,quartz_paragraph_question_plain_concat,kilt_tasks_hotpotqa_complex_question,quartz_use_info_from_paragraph_question | lora | | cluster_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/multi_news_1_0_0,para_crawl_enes,quoref_Context_Contains_Answer,cos_e_v1_11_description_question_option_id,glue_stsb_2_0_0,quail_context_description_question_answer_text,dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_description_option_text,cos_e_v1_11_description_question_option_text,race_high_Write_a_multi_choice_question_options_given_,ropes_prompt_bottom_no_hint,quarel_testing_students,wmt16_translate_tr_en_1_0_0,duorc_SelfRC_generate_question,wiqa_what_is_the_final_step_of_the_following_process,quoref_Given_Context_Answer_Question,wiki_hop_original_generate_object,quartz_use_info_from_question_paragraph,duorc_SelfRC_build_story_around_qa,drop_2_0_0,wiqa_effect_with_string_answer,race_high_Taking_a_test,wiki_hop_original_generate_subject_and_object,glue_qqp_2_0_0,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | cluster_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/snli_1_1_0,duorc_SelfRC_question_answering,cot_sensemaking_ii,huggingface_xsum,duorc_ParaphraseRC_build_story_around_qa,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,quartz_answer_question_based_on,ropes_plain_background_situation,race_middle_Write_a_multi_choice_question_for_the_following_article,quail_description_context_question_text,web_questions_whats_the_answer,cot_ecqa,true_case,adversarial_qa_dbert_question_context_answer,duorc_SelfRC_title_generation,quail_context_question_description_answer_id,quarel_do_not_use,adversarial_qa_dbidaf_answer_the_following_q,duorc_ParaphraseRC_decide_worth_it,race_middle_Is_this_the_right_answer,wmt16_translate_de_en_1_0_0,wiki_qa_Is_This_True_,race_middle_Select_the_best_answer,aeslc_1_0_0,duorc_ParaphraseRC_question_answering,wiki_hop_original_choose_best_object_affirmative_1 | lora | | cluster_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/quoref_Found_Context_Online,social_i_qa_Show_choices_and_generate_answer,cos_e_v1_11_explain_why_human,cot_sensemaking,quoref_Find_Answer,quail_context_description_question_text,social_i_qa_Generate_the_question_from_the_answer,quartz_having_read_above_passage,stream_qed_ii,wiqa_what_might_be_the_first_step_of_the_process,wiki_bio_who,duorc_SelfRC_generate_question_by_answer,race_high_Select_the_best_answer_generate_span_,lambada_1_0_0,coqa_1_0_0,race_high_Select_the_best_answer,adversarial_qa_droberta_question_context_answer,quail_context_question_answer_description_id,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_tell_what_it_is,cnn_dailymail_3_4_0,quail_no_prompt_id,ag_news_subset_1_0_0,trivia_qa_rc_1_1_0,ropes_prompt_bottom_hint_beginning,super_glue_wic_1_0_2 | lora | | cluster_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2,duorc_ParaphraseRC_movie_director,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,quartz_read_passage_below_choose,social_i_qa_Check_if_a_random_answer_is_valid_or_not,cot_creak,ropes_plain_bottom_hint,super_glue_copa_1_0_2,natural_questions_open_1_0_0,trec_1_0_0,gem_web_nlg_en_1_1_0,wiki_bio_key_content,wmt16_translate_fi_en_1_0_0,quoref_Answer_Question_Given_Context,duorc_ParaphraseRC_generate_question,math_dataset_algebra__linear_1d_1_0_0,duorc_ParaphraseRC_title_generation,quail_context_description_question_answer_id,wiki_bio_what_content,adversarial_qa_dbert_tell_what_it_is,sciq_Multiple_Choice_Closed_Book_,duorc_ParaphraseRC_extract_answer,dream_baseline,gem_wiki_lingua_english_en_1_1_0 | lora | | cluster_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id,social_i_qa_Show_choices_and_generate_index,web_questions_potential_correct_answer,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cosmos_qa_1_0_0,sciq_Direct_Question,super_glue_wsc_fixed_1_0_2,race_middle_Taking_a_test,wmt14_translate_fr_en_1_0_0,duorc_SelfRC_extract_answer,wiki_hop_original_generate_subject,duorc_SelfRC_answer_question,qasc_is_correct_1,cos_e_v1_11_i_think,wiki_qa_exercise,race_middle_Write_a_multi_choice_question_options_given_,quoref_Read_And_Extract_,web_questions_short_general_knowledge_q,web_questions_question_answer,quarel_logic_test,app_reviews_categorize_rating_using_review,cot_strategyqa_ii,glue_mnli_2_0_0,quoref_Answer_Test,super_glue_rte_1_0_2 | lora | | cluster_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/gem_dart_1_1_0,quarel_choose_between,wiki_hop_original_explain_relation,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,race_high_Select_the_best_answer_no_instructions_,kilt_tasks_hotpotqa_combining_facts,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_generate_question,glue_cola_2_0_0,imdb_reviews_plain_text_1_0_0,squad_v1_1_3_0_0,race_high_Is_this_the_right_answer,qasc_qa_with_separated_facts_1,glue_sst2_2_0_0,wiqa_what_is_the_missing_first_step,duorc_ParaphraseRC_generate_question_by_answer,wmt16_translate_ro_en_1_0_0,wiki_qa_Jeopardy_style,quartz_answer_question_below,ropes_prompt_beginning,ropes_read_background_situation,wiki_qa_Decide_good_answer,super_glue_cb_1_0_2,qasc_qa_with_separated_facts_4,cot_ecqa_ii | lora | | cluster_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cot_creak_ii,quail_description_context_question_answer_id,kilt_tasks_hotpotqa_final_exam,cot_gsm8k,cos_e_v1_11_aligned_with_common_sense,squad_v2_0_3_0_0,duorc_SelfRC_movie_director,anli_r2_0_1_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,duorc_ParaphraseRC_answer_question,ropes_background_situation_middle,dream_generate_first_utterance,quail_context_question_description_text,adversarial_qa_droberta_based_on,glue_wnli_2_0_0,super_glue_record_1_0_2,web_questions_get_the_answer,ropes_prompt_mix,app_reviews_generate_review,cos_e_v1_11_rationale,adversarial_qa_droberta_generate_question,yelp_polarity_reviews_0_2_0,ropes_given_background_situation,qasc_qa_with_separated_facts_3,quoref_Guess_Answer | lora | | cluster_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article,cot_qasc,definite_pronoun_resolution_1_1_0,wiki_qa_found_on_google,wiki_bio_comprehension,wiki_qa_Topic_Prediction_Question_Only,wiki_bio_guess_person,fix_punct,race_middle_Select_the_best_answer_no_instructions_,quac_1_0_0,wiqa_does_the_supposed_perturbation_have_an_effect,quartz_given_the_fact_answer_the_q,ropes_new_situation_background_answer,social_i_qa_Generate_answer,gigaword_1_2_0,duorc_SelfRC_decide_worth_it,kilt_tasks_hotpotqa_straighforward_qa,quail_no_prompt_text,cot_esnli,quoref_Answer_Friend_Question,race_middle_Select_the_best_answer_generate_span_,unified_qa_science_inst,sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_strategyqa | lora | | cluster_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/stream_qed,cot_esnli_ii,quarel_heres_a_story,quoref_Guess_Title_For_Context,qasc_is_correct_2,wiqa_effect_with_label_answer,dream_generate_last_utterance,adversarial_qa_dbert_based_on,dream_answer_to_dialogue,sciq_Multiple_Choice_Question_First,quail_context_question_description_answer_text,wiki_qa_Direct_Answer_to_Question,ropes_background_new_situation_answer,adversarial_qa_droberta_answer_the_following_q,kilt_tasks_hotpotqa_formulate,quoref_What_Is_The_Answer,dbpedia_14_given_a_choice_of_categories_,qasc_qa_with_separated_facts_2,glue_mrpc_2_0_0,gem_e2e_nlg_1_1_0,anli_r1_0_1_0,race_middle_Read_the_article_and_answer_the_question_no_option_,wiki_hop_original_choose_best_object_interrogative_1,quail_description_context_question_answer_text,adversarial_qa_dbidaf_based_on | lora | Last updated on: 2024-01-26T15:04:01.000Z
[ "SCIQ" ]
CyberHarem/leto_arknights
CyberHarem
text-to-image
[ "art", "not-for-all-audiences", "text-to-image", "dataset:Cyberharem/leto_arknights", "license:mit", "region:us" ]
"2024-01-26T12:04:30Z"
2024-03-23T00:17:49+00:00
0
0
--- datasets: - Cyberharem/leto_arknights license: mit pipeline_tag: text-to-image tags: - art - not-for-all-audiences --- # LoRA model of leto/烈夏 (Arknights) ## What Is This? This is the LoRA model of waifu leto/烈夏 (Arknights). ## How Is It Trained? * This model is trained with [kohya-ss/sd-scripts](https://github.com/kohya-ss/sd-scripts), and the test images are generated with [a1111's webui](AUTOMATIC1111/stable-diffusion-webui) and [API sdk](https://github.com/mix1009/sdwebuiapi). * The [auto-training framework](https://github.com/deepghs/cyberharem) is maintained by [DeepGHS Team](https://huggingface.co/deepghs). The architecture of base model is is `SD1.5`. * Dataset used for training is the `stage3-p480-1200` in [Cyberharem/leto_arknights](https://huggingface.co/datasets/Cyberharem/leto_arknights), which contains 83 images. * **Trigger word is `leto_arknights`.** * Pruned core tags for this waifu are `animal ears, bear ears, multicolored hair, streaked hair, brown hair, hair ornament, red eyes, short hair, black hair, white hair, long hair`. You can add them to the prompt when some features of waifu (e.g. hair color) are not stable. * For more details in training, you can take a look at [training configuration file](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/train.toml). * For more details in LoRA, you can download it, and read the metadata with a1111's webui. ## How to Use It? After downloading the safetensors files for the specified step, you need to use them like common LoRA. * Recommended LoRA weight is 0.5-0.85. * Recommended trigger word weight is 0.7-1.1. For example, if you want to use the model from step 2070, you need to download [`2070/leto_arknights.safetensors`](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/2070/leto_arknights.safetensors) as LoRA. By using this model, you can generate images for the desired characters. ## Which Step Should I Use? We selected 5 good steps for you to choose. The best one is step 2070. 1026 images (1.03 GiB) were generated for auto-testing. ![Metrics Plot](metrics_plot.png) The base model used for generating preview images is [meinamix_v11](https://huggingface.co/meinamix_v11). Here are the preview of the recommended steps: | Step | Epoch | CCIP | AI Corrupt | Bikini Plus | Score | Download | pattern_0_0 | pattern_0_1 | pattern_0_2 | portrait_0 | portrait_1 | portrait_2 | full_body_0 | full_body_1 | profile_0 | profile_1 | free_0 | free_1 | shorts | maid_0 | maid_1 | miko | yukata | suit | china | bikini_0 | bikini_1 | bikini_2 | sit | squat | kneel | jump | crossed_arms | angry | smile | cry | grin | n_lie_0 | n_lie_1 | n_stand_0 | n_stand_1 | n_stand_2 | n_sex_0 | n_sex_1 | |-------:|--------:|:----------|:-------------|:--------------|:----------|:--------------------------------------------------------------------------------------------------|:----------------------------------------------|:----------------------------------------------|:----------------------------------------------|:--------------------------------------------|:--------------------------------------------|:--------------------------------------------|:----------------------------------------------|:----------------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:------------------------------------|:--------------------------------|:------------------------------------|:--------------------------------|:----------------------------------|:----------------------------------------|:----------------------------------------|:----------------------------------------|:------------------------------|:----------------------------------|:----------------------------------|:--------------------------------|:------------------------------------------------|:----------------------------------|:----------------------------------|:------------------------------|:--------------------------------|:--------------------------------------|:--------------------------------------|:------------------------------------------|:------------------------------------------|:------------------------------------------|:--------------------------------------|:--------------------------------------| | 2070 | 69 | 0.917 | 0.984 | **0.862** | **0.681** | [Download](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/2070/leto_arknights.zip) | ![pattern_0_0](2070/previews/pattern_0_0.png) | ![pattern_0_1](2070/previews/pattern_0_1.png) | ![pattern_0_2](2070/previews/pattern_0_2.png) | ![portrait_0](2070/previews/portrait_0.png) | ![portrait_1](2070/previews/portrait_1.png) | ![portrait_2](2070/previews/portrait_2.png) | ![full_body_0](2070/previews/full_body_0.png) | ![full_body_1](2070/previews/full_body_1.png) | ![profile_0](2070/previews/profile_0.png) | ![profile_1](2070/previews/profile_1.png) | ![free_0](2070/previews/free_0.png) | ![free_1](2070/previews/free_1.png) | ![shorts](2070/previews/shorts.png) | ![maid_0](2070/previews/maid_0.png) | ![maid_1](2070/previews/maid_1.png) | ![miko](2070/previews/miko.png) | ![yukata](2070/previews/yukata.png) | ![suit](2070/previews/suit.png) | ![china](2070/previews/china.png) | ![bikini_0](2070/previews/bikini_0.png) | ![bikini_1](2070/previews/bikini_1.png) | ![bikini_2](2070/previews/bikini_2.png) | ![sit](2070/previews/sit.png) | ![squat](2070/previews/squat.png) | ![kneel](2070/previews/kneel.png) | ![jump](2070/previews/jump.png) | ![crossed_arms](2070/previews/crossed_arms.png) | ![angry](2070/previews/angry.png) | ![smile](2070/previews/smile.png) | ![cry](2070/previews/cry.png) | ![grin](2070/previews/grin.png) | ![n_lie_0](2070/previews/n_lie_0.png) | ![n_lie_1](2070/previews/n_lie_1.png) | ![n_stand_0](2070/previews/n_stand_0.png) | ![n_stand_1](2070/previews/n_stand_1.png) | ![n_stand_2](2070/previews/n_stand_2.png) | ![n_sex_0](2070/previews/n_sex_0.png) | ![n_sex_1](2070/previews/n_sex_1.png) | | 2400 | 80 | 0.917 | 0.989 | 0.858 | 0.675 | [Download](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/2400/leto_arknights.zip) | ![pattern_0_0](2400/previews/pattern_0_0.png) | ![pattern_0_1](2400/previews/pattern_0_1.png) | ![pattern_0_2](2400/previews/pattern_0_2.png) | ![portrait_0](2400/previews/portrait_0.png) | ![portrait_1](2400/previews/portrait_1.png) | ![portrait_2](2400/previews/portrait_2.png) | ![full_body_0](2400/previews/full_body_0.png) | ![full_body_1](2400/previews/full_body_1.png) | ![profile_0](2400/previews/profile_0.png) | ![profile_1](2400/previews/profile_1.png) | ![free_0](2400/previews/free_0.png) | ![free_1](2400/previews/free_1.png) | ![shorts](2400/previews/shorts.png) | ![maid_0](2400/previews/maid_0.png) | ![maid_1](2400/previews/maid_1.png) | ![miko](2400/previews/miko.png) | ![yukata](2400/previews/yukata.png) | ![suit](2400/previews/suit.png) | ![china](2400/previews/china.png) | ![bikini_0](2400/previews/bikini_0.png) | ![bikini_1](2400/previews/bikini_1.png) | ![bikini_2](2400/previews/bikini_2.png) | ![sit](2400/previews/sit.png) | ![squat](2400/previews/squat.png) | ![kneel](2400/previews/kneel.png) | ![jump](2400/previews/jump.png) | ![crossed_arms](2400/previews/crossed_arms.png) | ![angry](2400/previews/angry.png) | ![smile](2400/previews/smile.png) | ![cry](2400/previews/cry.png) | ![grin](2400/previews/grin.png) | ![n_lie_0](2400/previews/n_lie_0.png) | ![n_lie_1](2400/previews/n_lie_1.png) | ![n_stand_0](2400/previews/n_stand_0.png) | ![n_stand_1](2400/previews/n_stand_1.png) | ![n_stand_2](2400/previews/n_stand_2.png) | ![n_sex_0](2400/previews/n_sex_0.png) | ![n_sex_1](2400/previews/n_sex_1.png) | | 1170 | 39 | 0.913 | **0.991** | 0.858 | 0.671 | [Download](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/1170/leto_arknights.zip) | ![pattern_0_0](1170/previews/pattern_0_0.png) | ![pattern_0_1](1170/previews/pattern_0_1.png) | ![pattern_0_2](1170/previews/pattern_0_2.png) | ![portrait_0](1170/previews/portrait_0.png) | ![portrait_1](1170/previews/portrait_1.png) | ![portrait_2](1170/previews/portrait_2.png) | ![full_body_0](1170/previews/full_body_0.png) | ![full_body_1](1170/previews/full_body_1.png) | ![profile_0](1170/previews/profile_0.png) | ![profile_1](1170/previews/profile_1.png) | ![free_0](1170/previews/free_0.png) | ![free_1](1170/previews/free_1.png) | ![shorts](1170/previews/shorts.png) | ![maid_0](1170/previews/maid_0.png) | ![maid_1](1170/previews/maid_1.png) | ![miko](1170/previews/miko.png) | ![yukata](1170/previews/yukata.png) | ![suit](1170/previews/suit.png) | ![china](1170/previews/china.png) | ![bikini_0](1170/previews/bikini_0.png) | ![bikini_1](1170/previews/bikini_1.png) | ![bikini_2](1170/previews/bikini_2.png) | ![sit](1170/previews/sit.png) | ![squat](1170/previews/squat.png) | ![kneel](1170/previews/kneel.png) | ![jump](1170/previews/jump.png) | ![crossed_arms](1170/previews/crossed_arms.png) | ![angry](1170/previews/angry.png) | ![smile](1170/previews/smile.png) | ![cry](1170/previews/cry.png) | ![grin](1170/previews/grin.png) | ![n_lie_0](1170/previews/n_lie_0.png) | ![n_lie_1](1170/previews/n_lie_1.png) | ![n_stand_0](1170/previews/n_stand_0.png) | ![n_stand_1](1170/previews/n_stand_1.png) | ![n_stand_2](1170/previews/n_stand_2.png) | ![n_sex_0](1170/previews/n_sex_0.png) | ![n_sex_1](1170/previews/n_sex_1.png) | | 450 | 15 | 0.919 | 0.988 | 0.855 | 0.671 | [Download](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/450/leto_arknights.zip) | ![pattern_0_0](450/previews/pattern_0_0.png) | ![pattern_0_1](450/previews/pattern_0_1.png) | ![pattern_0_2](450/previews/pattern_0_2.png) | ![portrait_0](450/previews/portrait_0.png) | ![portrait_1](450/previews/portrait_1.png) | ![portrait_2](450/previews/portrait_2.png) | ![full_body_0](450/previews/full_body_0.png) | ![full_body_1](450/previews/full_body_1.png) | ![profile_0](450/previews/profile_0.png) | ![profile_1](450/previews/profile_1.png) | ![free_0](450/previews/free_0.png) | ![free_1](450/previews/free_1.png) | ![shorts](450/previews/shorts.png) | ![maid_0](450/previews/maid_0.png) | ![maid_1](450/previews/maid_1.png) | ![miko](450/previews/miko.png) | ![yukata](450/previews/yukata.png) | ![suit](450/previews/suit.png) | ![china](450/previews/china.png) | ![bikini_0](450/previews/bikini_0.png) | ![bikini_1](450/previews/bikini_1.png) | ![bikini_2](450/previews/bikini_2.png) | ![sit](450/previews/sit.png) | ![squat](450/previews/squat.png) | ![kneel](450/previews/kneel.png) | ![jump](450/previews/jump.png) | ![crossed_arms](450/previews/crossed_arms.png) | ![angry](450/previews/angry.png) | ![smile](450/previews/smile.png) | ![cry](450/previews/cry.png) | ![grin](450/previews/grin.png) | ![n_lie_0](450/previews/n_lie_0.png) | ![n_lie_1](450/previews/n_lie_1.png) | ![n_stand_0](450/previews/n_stand_0.png) | ![n_stand_1](450/previews/n_stand_1.png) | ![n_stand_2](450/previews/n_stand_2.png) | ![n_sex_0](450/previews/n_sex_0.png) | ![n_sex_1](450/previews/n_sex_1.png) | | 1080 | 36 | **0.921** | 0.990 | 0.851 | 0.666 | [Download](https://huggingface.co/Cyberharem/leto_arknights/resolve/main/1080/leto_arknights.zip) | ![pattern_0_0](1080/previews/pattern_0_0.png) | ![pattern_0_1](1080/previews/pattern_0_1.png) | ![pattern_0_2](1080/previews/pattern_0_2.png) | ![portrait_0](1080/previews/portrait_0.png) | ![portrait_1](1080/previews/portrait_1.png) | ![portrait_2](1080/previews/portrait_2.png) | ![full_body_0](1080/previews/full_body_0.png) | ![full_body_1](1080/previews/full_body_1.png) | ![profile_0](1080/previews/profile_0.png) | ![profile_1](1080/previews/profile_1.png) | ![free_0](1080/previews/free_0.png) | ![free_1](1080/previews/free_1.png) | ![shorts](1080/previews/shorts.png) | ![maid_0](1080/previews/maid_0.png) | ![maid_1](1080/previews/maid_1.png) | ![miko](1080/previews/miko.png) | ![yukata](1080/previews/yukata.png) | ![suit](1080/previews/suit.png) | ![china](1080/previews/china.png) | ![bikini_0](1080/previews/bikini_0.png) | ![bikini_1](1080/previews/bikini_1.png) | ![bikini_2](1080/previews/bikini_2.png) | ![sit](1080/previews/sit.png) | ![squat](1080/previews/squat.png) | ![kneel](1080/previews/kneel.png) | ![jump](1080/previews/jump.png) | ![crossed_arms](1080/previews/crossed_arms.png) | ![angry](1080/previews/angry.png) | ![smile](1080/previews/smile.png) | ![cry](1080/previews/cry.png) | ![grin](1080/previews/grin.png) | ![n_lie_0](1080/previews/n_lie_0.png) | ![n_lie_1](1080/previews/n_lie_1.png) | ![n_stand_0](1080/previews/n_stand_0.png) | ![n_stand_1](1080/previews/n_stand_1.png) | ![n_stand_2](1080/previews/n_stand_2.png) | ![n_sex_0](1080/previews/n_sex_0.png) | ![n_sex_1](1080/previews/n_sex_1.png) | ## Anything Else? Because the automation of LoRA training always annoys some people. So for the following groups, it is not recommended to use this model and we express regret: 1. Individuals who cannot tolerate any deviations from the original character design, even in the slightest detail. 2. Individuals who are facing the application scenarios with high demands for accuracy in recreating character outfits. 3. Individuals who cannot accept the potential randomness in AI-generated images based on the Stable Diffusion algorithm. 4. Individuals who are not comfortable with the fully automated process of training character models using LoRA, or those who believe that training character models must be done purely through manual operations to avoid disrespecting the characters. 5. Individuals who finds the generated image content offensive to their values. ## All Steps We uploaded the files in all steps. you can check the images, metrics and download them in the following links: * [Steps From 1620 to 2400](all/0.md) * [Steps From 720 to 1530](all/1.md) * [Steps From 90 to 630](all/2.md)
[ "BEAR" ]
zhan1993/gptneo_1B_flan_10_experts-epoch_0
zhan1993
null
[ "region:us" ]
"2024-01-26T22:17:58Z"
2024-01-27T14:20:24+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_10experts_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_record_1_0_2,quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quac_1_0_0,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | cluster_10experts_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,ropes_background_new_situation_answer,ropes_plain_background_situation,adversarial_qa_droberta_tell_what_it_is,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,adversarial_qa_droberta_based_on,adversarial_qa_dbert_answer_the_following_q,ropes_prompt_beginning,ropes_read_background_situation,adversarial_qa_droberta_question_context_answer,ropes_plain_bottom_hint,quoref_Answer_Question_Given_Context,squad_v2_0_3_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is,ropes_prompt_mix,ropes_background_situation_middle,squad_v1_1_3_0_0 | lora | | cluster_10experts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0,trec_1_0_0,cos_e_v1_11_rationale,natural_questions_open_1_0_0,web_questions_whats_the_answer,kilt_tasks_hotpotqa_final_exam,cos_e_v1_11_generate_explanation_given_text,cot_gsm8k_ii,web_questions_question_answer,stream_aqua,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,cos_e_v1_11_i_think,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_ecqa_ii,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,cos_e_v1_11_explain_why_human,cot_creak_ii,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,stream_qed_ii,wiki_bio_guess_person,cos_e_v1_11_aligned_with_common_sense,cot_sensemaking_ii,kilt_tasks_hotpotqa_formulate,social_i_qa_I_was_wondering | lora | | cluster_10experts_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2,wiki_qa_found_on_google,super_glue_wic_1_0_2,glue_cola_2_0_0,paws_wiki_1_1_0,super_glue_wsc_fixed_1_0_2,wiki_qa_Is_This_True_,snli_1_1_0,glue_qqp_2_0_0,wiki_qa_automatic_system,qasc_is_correct_1,glue_stsb_2_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,social_i_qa_Check_if_a_random_answer_is_valid_or_not,qasc_is_correct_2,glue_wnli_2_0_0,glue_mrpc_2_0_0 | lora | | cluster_10experts_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_sensemaking,yelp_polarity_reviews_0_2_0,anli_r3_0_1_0,lambada_1_0_0,wmt16_translate_ro_en_1_0_0,dream_generate_last_utterance,ag_news_subset_1_0_0,gem_dart_1_1_0,gem_common_gen_1_1_0,cot_creak,adversarial_qa_dbidaf_generate_question,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,gem_web_nlg_en_1_1_0,word_segment,anli_r2_0_1_0,app_reviews_generate_review,wmt16_translate_de_en_1_0_0,anli_r1_0_1_0,cot_ecqa,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,social_i_qa_Generate_the_question_from_the_answer,para_crawl_enes,race_high_Write_a_multi_choice_question_options_given_,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0,cosmos_qa_1_0_0,cot_esnli_ii | lora | | cluster_10experts_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_middle_Is_this_the_right_answer,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,race_high_Is_this_the_right_answer,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,dream_baseline,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | cluster_10experts_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,cot_gsm8k,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,cnn_dailymail_3_4_0,duorc_SelfRC_build_story_around_qa,gem_wiki_lingua_english_en_1_1_0,wiki_bio_what_content,race_high_Write_a_multi_choice_question_for_the_following_article,wiki_bio_who,aeslc_1_0_0,dream_answer_to_dialogue,gem_e2e_nlg_1_1_0,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | cluster_10experts_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,cos_e_v1_11_question_option_description_id,quarel_do_not_use,qasc_qa_with_separated_facts_5,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,quarel_heres_a_story,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,ropes_plain_no_background,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,quarel_testing_students,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,social_i_qa_Generate_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | cluster_10experts_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,app_reviews_categorize_rating_using_review,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,wiqa_does_the_supposed_perturbation_have_an_effect,math_dataset_algebra__linear_1d_1_0_0,unified_qa_science_inst,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cot_strategyqa,wiqa_effect_with_string_answer,wiki_qa_exercise,wiki_qa_Topic_Prediction_Question_Only,super_glue_copa_1_0_2,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,cot_qasc,cot_strategyqa_ii,stream_aqua_ii | lora | | cluster_10experts_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_object,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_explain_relation,wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_generate_subject,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object | lora | Last updated on: 2024-01-27T09:13:37.000Z
[ "SCIQ" ]
zhan1993/gptneo_1B_flan_10_experts-epoch_1
zhan1993
null
[ "region:us" ]
"2024-01-26T23:47:52Z"
2024-01-27T15:23:21+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_10experts_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_record_1_0_2,quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quac_1_0_0,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | cluster_10experts_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,ropes_background_new_situation_answer,ropes_plain_background_situation,adversarial_qa_droberta_tell_what_it_is,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,adversarial_qa_droberta_based_on,adversarial_qa_dbert_answer_the_following_q,ropes_prompt_beginning,ropes_read_background_situation,adversarial_qa_droberta_question_context_answer,ropes_plain_bottom_hint,quoref_Answer_Question_Given_Context,squad_v2_0_3_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is,ropes_prompt_mix,ropes_background_situation_middle,squad_v1_1_3_0_0 | lora | | cluster_10experts_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0,trec_1_0_0,cos_e_v1_11_rationale,natural_questions_open_1_0_0,web_questions_whats_the_answer,kilt_tasks_hotpotqa_final_exam,cos_e_v1_11_generate_explanation_given_text,cot_gsm8k_ii,web_questions_question_answer,stream_aqua,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,cos_e_v1_11_i_think,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_ecqa_ii,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,cos_e_v1_11_explain_why_human,cot_creak_ii,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,stream_qed_ii,wiki_bio_guess_person,cos_e_v1_11_aligned_with_common_sense,cot_sensemaking_ii,kilt_tasks_hotpotqa_formulate,social_i_qa_I_was_wondering | lora | | cluster_10experts_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2,wiki_qa_found_on_google,super_glue_wic_1_0_2,glue_cola_2_0_0,paws_wiki_1_1_0,super_glue_wsc_fixed_1_0_2,wiki_qa_Is_This_True_,snli_1_1_0,glue_qqp_2_0_0,wiki_qa_automatic_system,qasc_is_correct_1,glue_stsb_2_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,social_i_qa_Check_if_a_random_answer_is_valid_or_not,qasc_is_correct_2,glue_wnli_2_0_0,glue_mrpc_2_0_0 | lora | | cluster_10experts_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_sensemaking,yelp_polarity_reviews_0_2_0,anli_r3_0_1_0,lambada_1_0_0,wmt16_translate_ro_en_1_0_0,dream_generate_last_utterance,ag_news_subset_1_0_0,gem_dart_1_1_0,gem_common_gen_1_1_0,cot_creak,adversarial_qa_dbidaf_generate_question,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,gem_web_nlg_en_1_1_0,word_segment,anli_r2_0_1_0,app_reviews_generate_review,wmt16_translate_de_en_1_0_0,anli_r1_0_1_0,cot_ecqa,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,social_i_qa_Generate_the_question_from_the_answer,para_crawl_enes,race_high_Write_a_multi_choice_question_options_given_,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0,cosmos_qa_1_0_0,cot_esnli_ii | lora | | cluster_10experts_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_middle_Is_this_the_right_answer,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,race_high_Is_this_the_right_answer,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,dream_baseline,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | cluster_10experts_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,cot_gsm8k,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,cnn_dailymail_3_4_0,duorc_SelfRC_build_story_around_qa,gem_wiki_lingua_english_en_1_1_0,wiki_bio_what_content,race_high_Write_a_multi_choice_question_for_the_following_article,wiki_bio_who,aeslc_1_0_0,dream_answer_to_dialogue,gem_e2e_nlg_1_1_0,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | cluster_10experts_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,cos_e_v1_11_question_option_description_id,quarel_do_not_use,qasc_qa_with_separated_facts_5,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,quarel_heres_a_story,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,ropes_plain_no_background,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,quarel_testing_students,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,social_i_qa_Generate_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | cluster_10experts_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,app_reviews_categorize_rating_using_review,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,wiqa_does_the_supposed_perturbation_have_an_effect,math_dataset_algebra__linear_1d_1_0_0,unified_qa_science_inst,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cot_strategyqa,wiqa_effect_with_string_answer,wiki_qa_exercise,wiki_qa_Topic_Prediction_Question_Only,super_glue_copa_1_0_2,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,cot_qasc,cot_strategyqa_ii,stream_aqua_ii | lora | | cluster_10experts_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_object,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_explain_relation,wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_generate_subject,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object | lora | Last updated on: 2024-01-27T11:57:47.000Z
[ "SCIQ" ]
zhan1993/gptneo_1B_flan_10_experts-epoch_2
zhan1993
null
[ "region:us" ]
"2024-01-27T21:46:31Z"
2024-03-06T22:03:09+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_sensemaking,yelp_polarity_reviews_0_2_0,anli_r3_0_1_0,lambada_1_0_0,wmt16_translate_ro_en_1_0_0,dream_generate_last_utterance,ag_news_subset_1_0_0,gem_dart_1_1_0,gem_common_gen_1_1_0,cot_creak,adversarial_qa_dbidaf_generate_question,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,gem_web_nlg_en_1_1_0,word_segment,anli_r2_0_1_0,app_reviews_generate_review,wmt16_translate_de_en_1_0_0,anli_r1_0_1_0,cot_ecqa,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,social_i_qa_Generate_the_question_from_the_answer,para_crawl_enes,race_high_Write_a_multi_choice_question_options_given_,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0,cosmos_qa_1_0_0,cot_esnli_ii | lora | | cluster_6 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_rte_1_0_2,wiki_qa_found_on_google,super_glue_wic_1_0_2,glue_cola_2_0_0,paws_wiki_1_1_0,super_glue_wsc_fixed_1_0_2,wiki_qa_Is_This_True_,snli_1_1_0,glue_qqp_2_0_0,wiki_qa_automatic_system,qasc_is_correct_1,glue_stsb_2_0_0,glue_qnli_2_0_0,glue_mnli_2_0_0,social_i_qa_Check_if_a_random_answer_is_valid_or_not,qasc_is_correct_2,glue_wnli_2_0_0,glue_mrpc_2_0_0 | lora | | cluster_5 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating,app_reviews_categorize_rating_using_review,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,wiqa_does_the_supposed_perturbation_have_an_effect,math_dataset_algebra__linear_1d_1_0_0,unified_qa_science_inst,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cot_strategyqa,wiqa_effect_with_string_answer,wiki_qa_exercise,wiki_qa_Topic_Prediction_Question_Only,super_glue_copa_1_0_2,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,cot_qasc,cot_strategyqa_ii,stream_aqua_ii | lora | | cluster_4 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,ropes_background_new_situation_answer,ropes_plain_background_situation,adversarial_qa_droberta_tell_what_it_is,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,adversarial_qa_droberta_based_on,adversarial_qa_dbert_answer_the_following_q,ropes_prompt_beginning,ropes_read_background_situation,adversarial_qa_droberta_question_context_answer,ropes_plain_bottom_hint,quoref_Answer_Question_Given_Context,squad_v2_0_3_0_0,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is,ropes_prompt_mix,ropes_background_situation_middle,squad_v1_1_3_0_0 | lora | | cluster_8 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,cos_e_v1_11_question_option_description_id,quarel_do_not_use,qasc_qa_with_separated_facts_5,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,quarel_heres_a_story,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,ropes_plain_no_background,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,quarel_testing_students,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,social_i_qa_Generate_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | cluster_2 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_middle_Is_this_the_right_answer,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,race_high_Is_this_the_right_answer,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,dream_baseline,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | cluster_1 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/glue_sst2_2_0_0,trec_1_0_0,cos_e_v1_11_rationale,natural_questions_open_1_0_0,web_questions_whats_the_answer,kilt_tasks_hotpotqa_final_exam,cos_e_v1_11_generate_explanation_given_text,cot_gsm8k_ii,web_questions_question_answer,stream_aqua,dbpedia_14_pick_one_category_for_the_following_text,dbpedia_14_given_a_choice_of_categories_,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,cos_e_v1_11_i_think,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,cot_ecqa_ii,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,cos_e_v1_11_explain_why_human,cot_creak_ii,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,stream_qed_ii,wiki_bio_guess_person,cos_e_v1_11_aligned_with_common_sense,cot_sensemaking_ii,kilt_tasks_hotpotqa_formulate,social_i_qa_I_was_wondering | lora | | cluster_7 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,cot_gsm8k,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,cnn_dailymail_3_4_0,duorc_SelfRC_build_story_around_qa,gem_wiki_lingua_english_en_1_1_0,wiki_bio_what_content,race_high_Write_a_multi_choice_question_for_the_following_article,wiki_bio_who,aeslc_1_0_0,dream_answer_to_dialogue,gem_e2e_nlg_1_1_0,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | cluster_9 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/super_glue_record_1_0_2,quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quac_1_0_0,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | cluster_10 | EleutherAI/gpt-neo-1.3B | sordonia/flan-10k-flat/wiki_hop_original_generate_object,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_explain_relation,wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_generate_subject,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object | lora | Last updated on: 2024-01-27T21:49:34.000Z
[ "SCIQ" ]
zhan1993/phi2_flan_10_random_unbalanced-epoch_0
zhan1993
null
[ "region:us" ]
"2024-01-31T11:45:24Z"
2024-02-08T04:28:58+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | cluster_3 | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,race_high_Write_a_multi_choice_question_for_the_following_article,kilt_tasks_hotpotqa_complex_question,quoref_Answer_Test,adversarial_qa_dbidaf_based_on,duorc_ParaphraseRC_title_generation,cot_strategyqa,sciq_Direct_Question,adversarial_qa_dbert_answer_the_following_q,quartz_paragraph_question_plain_concat,wiki_hop_original_generate_subject,race_middle_Read_the_article_and_answer_the_question_no_option_,adversarial_qa_dbert_tell_what_it_is,cos_e_v1_11_aligned_with_common_sense,anli_r2_0_1_0,cot_gsm8k,qasc_qa_with_separated_facts_1,wiqa_effect_with_string_answer,wiki_bio_what_content,cot_qasc,gem_dart_1_1_0,natural_questions_open_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,wiki_qa_Topic_Prediction_Answer_Only,dream_generate_first_utterance,dream_read_the_following_conversation_and_answer_the_question,ropes_prompt_mix,wmt16_translate_ro_en_1_0_0,gem_wiki_lingua_english_en_1_1_0,social_i_qa_Generate_answer,cot_gsm8k_ii,stream_aqua_ii,quoref_Context_Contains_Answer,quail_context_question_description_text,ropes_prompt_beginning,drop_2_0_0 | lora | | cluster_6 | phi-2 | sordonia/flan-10k-flat/stream_aqua,anli_r3_0_1_0,quail_context_question_description_answer_id,wiki_hop_original_choose_best_object_interrogative_1,true_case,wmt16_translate_tr_en_1_0_0,qasc_is_correct_1,ropes_prompt_bottom_hint_beginning,quarel_heres_a_story,wiki_hop_original_explain_relation,adversarial_qa_droberta_answer_the_following_q,lambada_1_0_0,squad_v2_0_3_0_0,wiqa_effect_with_label_answer,cos_e_v1_11_i_think,quoref_Guess_Answer,cos_e_v1_11_question_description_option_text,wiki_qa_found_on_google,duorc_SelfRC_build_story_around_qa,quartz_read_passage_below_choose,qasc_qa_with_separated_facts_2,wiqa_what_might_be_the_last_step_of_the_process,multi_news_1_0_0,quoref_Read_And_Extract_,adversarial_qa_dbert_question_context_answer,app_reviews_categorize_rating_using_review,qasc_is_correct_2 | lora | | cluster_5 | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer,duorc_SelfRC_answer_question,wiki_qa_Is_This_True_,cos_e_v1_11_explain_why_human,race_middle_Write_a_multi_choice_question_options_given_,dbpedia_14_pick_one_category_for_the_following_text,quartz_answer_question_based_on,fix_punct,squad_v1_1_3_0_0,sciq_Multiple_Choice_Question_First,quoref_Given_Context_Answer_Question,super_glue_copa_1_0_2,cnn_dailymail_3_4_0,race_middle_Is_this_the_right_answer,quail_context_description_question_answer_text,race_high_Read_the_article_and_answer_the_question_no_option_,duorc_ParaphraseRC_generate_question_by_answer,imdb_reviews_plain_text_1_0_0,quartz_use_info_from_question_paragraph,ropes_plain_bottom_hint,quarel_choose_between,glue_sst2_2_0_0,adversarial_qa_dbert_based_on,wmt16_translate_de_en_1_0_0 | lora | | cluster_4 | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id,duorc_ParaphraseRC_movie_director,super_glue_cb_1_0_2,wiqa_what_is_the_final_step_of_the_following_process,cot_creak,glue_mnli_2_0_0,wiki_qa_Topic_Prediction_Question_Only,quarel_logic_test,ropes_plain_no_background,gem_e2e_nlg_1_1_0,wiqa_does_the_supposed_perturbation_have_an_effect,cot_ecqa,quarel_testing_students,wiki_bio_comprehension,wmt14_translate_fr_en_1_0_0,cos_e_v1_11_description_question_option_text,social_i_qa_Check_if_a_random_answer_is_valid_or_not,duorc_ParaphraseRC_decide_worth_it | lora | | cluster_8 | phi-2 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q,dream_generate_last_utterance,web_questions_question_answer,quoref_What_Is_The_Answer,stream_qed_ii,cos_e_v1_11_question_description_option_id,adversarial_qa_droberta_based_on,para_crawl_enes,glue_qqp_2_0_0,cos_e_v1_11_generate_explanation_given_text,glue_mrpc_2_0_0,duorc_ParaphraseRC_answer_question | lora | | cluster_2 | phi-2 | sordonia/flan-10k-flat/ropes_read_background_situation,yelp_polarity_reviews_0_2_0,kilt_tasks_hotpotqa_straighforward_qa,qasc_qa_with_combined_facts_1,snli_1_1_0,wiki_hop_original_choose_best_object_affirmative_2,cot_strategyqa_ii,gem_common_gen_1_1_0,race_middle_Select_the_best_answer_no_instructions_,quoref_Find_Answer,trec_1_0_0,duorc_SelfRC_question_answering,race_middle_Taking_a_test,ropes_plain_background_situation,super_glue_record_1_0_2,ropes_background_new_situation_answer,cos_e_v1_11_rationale,web_questions_get_the_answer,quail_no_prompt_id,quoref_Answer_Question_Given_Context,duorc_SelfRC_movie_director,app_reviews_convert_to_star_rating,duorc_SelfRC_decide_worth_it,stream_qed | lora | | cluster_1 | phi-2 | sordonia/flan-10k-flat/anli_r1_0_1_0,dream_answer_to_dialogue,wiki_bio_guess_person,web_questions_potential_correct_answer,ropes_new_situation_background_answer,duorc_SelfRC_title_generation,quartz_use_info_from_paragraph_question,quartz_having_read_above_passage,super_glue_wic_1_0_2,huggingface_xsum,cot_ecqa_ii,cos_e_v1_11_question_option_description_id,race_middle_Select_the_best_answer,kilt_tasks_hotpotqa_combining_facts,cot_creak_ii,race_high_Select_the_best_answer_no_instructions_,sciq_Multiple_Choice | lora | | cluster_7 | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer,adversarial_qa_dbidaf_generate_question,math_dataset_algebra__linear_1d_1_0_0,race_high_Is_this_the_right_answer,adversarial_qa_droberta_generate_question,wiqa_which_of_the_following_is_the_supposed_perturbation,adversarial_qa_droberta_question_context_answer,ag_news_subset_1_0_0,adversarial_qa_droberta_tell_what_it_is,wiki_qa_automatic_system,quail_description_context_question_answer_text,web_questions_whats_the_answer,gem_web_nlg_en_1_1_0,ropes_given_background_situation,ropes_background_situation_middle,quartz_answer_question_below,duorc_ParaphraseRC_extract_answer,social_i_qa_I_was_wondering,wiki_hop_original_choose_best_object_affirmative_1,cos_e_v1_11_question_option_description_text,coqa_1_0_0,qasc_qa_with_separated_facts_4,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,app_reviews_generate_review,wiki_qa_Direct_Answer_to_Question,race_middle_Select_the_best_answer_generate_span_,race_high_Taking_a_test,glue_stsb_2_0_0,wiki_qa_Decide_good_answer,super_glue_wsc_fixed_1_0_2,social_i_qa_Show_choices_and_generate_answer,adversarial_qa_dbidaf_question_context_answer,cot_esnli,cot_esnli_ii,wiki_qa_Jeopardy_style,quoref_Answer_Friend_Question | lora | | cluster_9 | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint,qasc_qa_with_separated_facts_5,quail_context_description_question_answer_id,wmt16_translate_fi_en_1_0_0,wiki_hop_original_generate_object,quoref_Guess_Title_For_Context,qasc_qa_with_separated_facts_3,wiki_qa_Generate_Question_from_Topic,duorc_ParaphraseRC_question_answering,social_i_qa_Show_choices_and_generate_index,quac_1_0_0,duorc_SelfRC_generate_question,kilt_tasks_hotpotqa_formulate,definite_pronoun_resolution_1_1_0,adversarial_qa_dbidaf_answer_the_following_q,dbpedia_14_given_a_choice_of_categories_,gigaword_1_2_0,race_high_Write_a_multi_choice_question_options_given_,quoref_Found_Context_Online,quail_context_question_description_answer_text,wiki_bio_key_content,quail_context_description_question_text,quail_context_question_answer_description_id | lora | | cluster_10 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object,aeslc_1_0_0,sciq_Multiple_Choice_Closed_Book_,wiki_qa_exercise,duorc_ParaphraseRC_build_story_around_qa,glue_wnli_2_0_0,wiki_hop_original_choose_best_object_interrogative_2,trivia_qa_rc_1_1_0,duorc_SelfRC_extract_answer,cot_sensemaking_ii,quail_description_context_question_text,quail_no_prompt_text,duorc_ParaphraseRC_generate_question,unified_qa_science_inst,app_reviews_convert_to_rating,glue_qnli_2_0_0,adversarial_qa_dbert_generate_question,wiki_bio_who,quail_description_context_question_answer_id,wiqa_what_might_be_the_first_step_of_the_process,dream_baseline,web_questions_short_general_knowledge_q,quarel_do_not_use,glue_cola_2_0_0,word_segment,race_high_Select_the_best_answer_generate_span_,duorc_SelfRC_generate_question_by_answer,cot_sensemaking,wiki_hop_original_choose_best_object_affirmative_3,super_glue_multirc_1_0_2,adversarial_qa_dbidaf_tell_what_it_is,paws_wiki_1_1_0,wiqa_what_is_the_missing_first_step,super_glue_rte_1_0_2,kilt_tasks_hotpotqa_final_exam,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,cosmos_qa_1_0_0,quail_context_question_answer_description_text | lora | Last updated on: 2024-02-01 08:00:01+00:00
[ "SCIQ" ]
ostapeno/phi2_10clusters_kmeans_balanced_1epoch
ostapeno
null
[ "region:us" ]
"2024-01-31T14:36:49Z"
2024-01-31T14:38:16+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_10c9o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google,super_glue_cb_1_0_2,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,unified_qa_science_inst,super_glue_multirc_1_0_2,cot_strategyqa,cot_ecqa_ii,quarel_do_not_use,wiki_qa_exercise,stream_qed_ii,super_glue_copa_1_0_2,social_i_qa_Generate_the_question_from_the_answer,quoref_Answer_Friend_Question,quarel_testing_students,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,glue_wnli_2_0_0,cot_qasc,cot_strategyqa_ii,stream_aqua_ii | lora | | phi2_joint_10c6o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,quartz_answer_question_based_on,ropes_plain_background_situation,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,wiqa_does_the_supposed_perturbation_have_an_effect,duorc_ParaphraseRC_title_generation,quartz_use_info_from_question_paragraph,wiqa_which_of_the_following_is_the_supposed_perturbation,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,wiqa_effect_with_string_answer,ropes_prompt_beginning,quartz_having_read_above_passage,ropes_read_background_situation,ropes_plain_bottom_hint,quarel_heres_a_story,ropes_plain_no_background,quartz_use_info_from_paragraph_question,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_10c5o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2,app_reviews_categorize_rating_using_review,cot_sensemaking,super_glue_wic_1_0_2,anli_r3_0_1_0,paws_wiki_1_1_0,cot_creak,wiki_qa_Is_This_True_,snli_1_1_0,glue_qqp_2_0_0,wiki_qa_automatic_system,cot_creak_ii,anli_r2_0_1_0,qasc_is_correct_1,anli_r1_0_1_0,glue_stsb_2_0_0,glue_qnli_2_0_0,cot_sensemaking_ii,glue_mnli_2_0_0,social_i_qa_Check_if_a_random_answer_is_valid_or_not,social_i_qa_I_was_wondering,qasc_is_correct_2,cosmos_qa_1_0_0,glue_mrpc_2_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_10c7o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question,app_reviews_convert_to_star_rating,true_case,natural_questions_open_1_0_0,web_questions_whats_the_answer,wiki_qa_Topic_Prediction_Answer_Only,kilt_tasks_hotpotqa_final_exam,web_questions_question_answer,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,app_reviews_convert_to_rating,adversarial_qa_droberta_based_on,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,squad_v2_0_3_0_0,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,wiki_qa_Topic_Prediction_Question_Only,kilt_tasks_hotpotqa_formulate,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,drop_2_0_0,squad_v1_1_3_0_0 | lora | | phi2_joint_10c1o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_middle_Is_this_the_right_answer,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,race_high_Is_this_the_right_answer,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | phi2_joint_10c8o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,stream_qed,super_glue_record_1_0_2,cos_e_v1_11_rationale,cos_e_v1_11_generate_explanation_given_text,wiki_hop_original_generate_object,adversarial_qa_droberta_tell_what_it_is,dbpedia_14_given_a_choice_of_categories_,wiki_hop_original_choose_best_object_affirmative_3,cos_e_v1_11_i_think,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_explain_relation,wiki_hop_original_choose_best_object_affirmative_1,adversarial_qa_dbert_answer_the_following_q,wiki_hop_original_choose_best_object_interrogative_2,adversarial_qa_droberta_question_context_answer,wiki_hop_original_generate_subject,cos_e_v1_11_aligned_with_common_sense,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | phi2_joint_10c0o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,huggingface_xsum,trec_1_0_0,yelp_polarity_reviews_0_2_0,lambada_1_0_0,dream_generate_last_utterance,ag_news_subset_1_0_0,wiki_bio_key_content,math_dataset_algebra__linear_1d_1_0_0,stream_aqua,dbpedia_14_pick_one_category_for_the_following_text,cnn_dailymail_3_4_0,duorc_SelfRC_build_story_around_qa,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,duorc_SelfRC_title_generation,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,wiki_bio_what_content,race_high_Write_a_multi_choice_question_for_the_following_article,wiki_bio_guess_person,cot_ecqa,dream_generate_first_utterance,dream_answer_to_dialogue,race_high_Write_a_multi_choice_question_options_given_,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | phi2_joint_10c2o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quac_1_0_0,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,duorc_SelfRC_extract_answer,quoref_Guess_Answer,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | | phi2_joint_10c3o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,sciq_Multiple_Choice,cos_e_v1_11_question_description_option_text,sciq_Direct_Question,qasc_qa_with_separated_facts_2,cos_e_v1_11_question_option_description_id,qasc_qa_with_separated_facts_5,dream_baseline,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_explain_why_human,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,cos_e_v1_11_description_question_option_id,quarel_choose_between,social_i_qa_Show_choices_and_generate_index,sciq_Multiple_Choice_Closed_Book_,qasc_qa_with_separated_facts_4,wiqa_effect_with_label_answer,quarel_logic_test,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_10c4o10_2e_1epoch_balanced_kmeans_on_sim_matrix | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,cot_esnli,cot_gsm8k,wmt16_translate_ro_en_1_0_0,glue_cola_2_0_0,gem_dart_1_1_0,wiqa_what_might_be_the_last_step_of_the_process,gem_common_gen_1_1_0,gem_wiki_lingua_english_en_1_1_0,fix_punct,gigaword_1_2_0,gem_web_nlg_en_1_1_0,word_segment,app_reviews_generate_review,wmt16_translate_de_en_1_0_0,wiki_bio_who,aeslc_1_0_0,wmt16_translate_fi_en_1_0_0,gem_e2e_nlg_1_1_0,para_crawl_enes,wmt14_translate_fr_en_1_0_0,cot_esnli_ii,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step | lora | Last updated on: 2024-01-31 14:36:51+00:00 Balanced k-means clusters created by applying k-means to similarity matrix (so task similarities are the features here)
[ "SCIQ" ]
OctaSpace/Mistral7B-fintuned-multi_x_science
OctaSpace
null
[ "safetensors", "arxiv:2010.14235", "license:apache-2.0", "region:us" ]
"2024-02-01T15:07:22Z"
2024-02-13T09:27:33+00:00
0
0
--- license: apache-2.0 --- # Fine-tuned Mistral Model for Multi-Document Summarization This model a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on [multi_x_science_sum](https://huggingface.co/datasets/multi_x_science_sum) dataset. ## Model description Mistral-7B-multixscience-finetuned is finetuned on multi_x_science_sum dataset in order to extend the capabilities of the original Mistral model in multi-document summarization tasks. The fine-tuned model leverages the power of Mistral model fundation, adapting it to synthesize and summarize information from multiple documents efficiently. ## Training and evaluation dataset Multi_x_science_sum is a large-scale multi-document summarization dataset created from scientific articles. Multi-XScience introduces a challenging multi-document summarization task: writing the related-work section of a paper based on its abstract and the articles it references. * [paper](https://arxiv.org/pdf/2010.14235.pdf) * [Source](https://huggingface.co/datasets/multi_x_science_sum) The training and evaluation datasets were uniquely generated to facilitate the fine-tuning of the model for multi-document summarization, particularly focusing on generating related work sections for scientific papers. Using a custom-designed prompt-generation process, the dataset is created to simulate the task of synthesizing related work sections based on a given paper's abstract and the abstracts of its referenced papers. ### Dataset Generation process The process involves generating prompts that instruct the model to use the abstract of the current paper along with the abstracts of cited papers to generate a new related work section. This approach aims to mimic the real-world scenario where a researcher synthesizes information from multiple sources to draft the related work section of a paper. * **Prompt Structure:** Each data point consists of an instructional prompt that includes: * The abstract of the current paper. * Abstracts from cited papers, labeled with unique identifiers. * An expected model response in the form of a generated related work section. ### Prompt generation Code ``` def generate_related_work_prompt(data): prompt = "[INST] <<SYS>>\n" prompt += "Use the abstract of the current paper and the abstracts of the cited papers to generate new related work.\n" prompt += "<</SYS>>\n\n" prompt += "Input:\nCurrent Paper's Abstract:\n- {}\n\n".format(data['abstract']) prompt += "Cited Papers' Abstracts:\n" for cite_id, cite_abstract in zip(data['ref_abstract']['cite_N'], data['ref_abstract']['abstract']): prompt += "- {}: {}\n".format(cite_id, cite_abstract) prompt += "\n[/INST]\n\nGenerated Related Work:\n{}\n".format(data['related_work']) return {"text": prompt} ``` The dataset generated through this process was used to train and evaluate the finetuned model, ensuring that it learns to accurately synthesize information from multiple sources into cohesive summaries. ## Training hyperparameters The following hyperparameters were used during training: ``` learning_rate: 2e-5 train_batch_size: 4 eval_batch_size: 4 seed: 42 optimizer: adamw_8bit num_epochs: 5 ``` ## Usage ``` import torch from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftConfig, PeftModel base_model = "mistralai/Mistral-7B-v0.1" adapter = "OctaSpace/Mistral7B-fintuned" # Load tokenizer tokenizer = AutoTokenizer.from_pretrained( base_model, add_bos_token=True, trust_remote_code=True, padding_side='left' ) # Create peft model using base_model and finetuned adapter config = PeftConfig.from_pretrained(adapter) model = AutoModelForCausalLM.from_pretrained(config.base_model_name_or_path, load_in_4bit=True, device_map='auto', torch_dtype='auto') model = PeftModel.from_pretrained(model, adapter) device = "cuda" if torch.cuda.is_available() else "cpu" model.to(device) model.eval() # Prompt content: messages = [] # Put here your related work generation instruction input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt').to(device) summary_ids = model.generate(input_ids=input_ids, max_new_tokens=512, do_sample=True, pad_token_id=2) summaries = tokenizer.batch_decode(summary_ids.detach().cpu().numpy(), skip_special_tokens = True) # Model response: print(summaries[0]) ```
[ "MULTI-XSCIENCE" ]
croissantllm/CroissantLLM_training_logs
croissantllm
null
[ "tensorboard", "legal", "code", "text-generation-inference", "art", "fr", "en", "dataset:cerebras/SlimPajama-627B", "dataset:uonlp/CulturaX", "dataset:pg19", "dataset:bigcode/starcoderdata", "dataset:manu/croissant_dataset", "arxiv:2402.00786", "license:mit", "region:us" ]
"2024-02-03T16:21:57Z"
2024-02-03T16:26:32+00:00
0
0
--- datasets: - cerebras/SlimPajama-627B - uonlp/CulturaX - pg19 - bigcode/starcoderdata - manu/croissant_dataset language: - fr - en license: mit tags: - legal - code - text-generation-inference - art --- # CroissantLLM - Base Training Logs These logs are the training tensorboard logs for the CroissantLLM initiative, and corresponds to the checkpoint after 190k steps (2.99 T) tokens. To play with the final model, we recommend using the Chat version: https://huggingface.co/croissantllm/CroissantLLMChat-v0.1. https://arxiv.org/abs/2402.00786 ## Abstract We introduce CroissantLLM, a 1.3B language model pretrained on a set of 3T English and French tokens, to bring to the research and industrial community a high-performance, fully open-sourced bilingual model that runs swiftly on consumer-grade local hardware. To that end, we pioneer the approach of training an intrinsically bilingual model with a 1:1 English-to-French pretraining data ratio, a custom tokenizer, and bilingual finetuning datasets. We release the training dataset, notably containing a French split with manually curated, high-quality, and varied data sources. To assess performance outside of English, we craft a novel benchmark, FrenchBench, consisting of an array of classification and generation tasks, covering various orthogonal aspects of model performance in the French Language. Additionally, rooted in transparency and to foster further Large Language Model research, we release codebases, and dozens of checkpoints across various model sizes, training data distributions, and training steps, as well as fine-tuned Chat models, and strong translation models. We evaluate our model through the FMTI framework, and validate 81% of the transparency criteria, far beyond the scores of even most open initiatives. This work enriches the NLP landscape, breaking away from previous English-centric work in order to strengthen our understanding of multilinguality in language models. ## Citation Our work can be cited as: ```bash @misc{faysse2024croissantllm, title={CroissantLLM: A Truly Bilingual French-English Language Model}, author={Manuel Faysse and Patrick Fernandes and Nuno Guerreiro and António Loison and Duarte Alves and Caio Corro and Nicolas Boizard and João Alves and Ricardo Rei and Pedro Martins and Antoni Bigata Casademunt and François Yvon and André Martins and Gautier Viaud and Céline Hudelot and Pierre Colombo}, year={2024}, eprint={2402.00786}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
[ "CRAFT" ]
biswa921/bge-m3
biswa921
null
[ "mteb", "model-index", "region:us" ]
"2024-02-06T05:39:28Z"
2024-03-22T12:51:52+00:00
0
0
--- tags: - mteb model-index: - name: bge-m3 results: - task: type: Classification dataset: name: MTEB AmazonCounterfactualClassification (en) type: mteb/amazon_counterfactual config: en split: test revision: e8379541af4e31359cca9fbcf4b00f2671dba205 metrics: - type: accuracy value: 75.6268656716418 - type: ap value: 39.50276109614102 - type: f1 value: 70.00224623431103 - task: type: Classification dataset: name: MTEB AmazonPolarityClassification type: mteb/amazon_polarity config: default split: test revision: e2d317d38cd51312af73b3d32a06d1a08b442046 metrics: - type: accuracy value: 91.013675 - type: ap value: 87.30227544778319 - type: f1 value: 91.00157923673694 - task: type: Classification dataset: name: MTEB AmazonReviewsClassification (en) type: mteb/amazon_reviews_multi config: en split: test revision: 1399c76144fd37290681b995c656ef9b2e06e26d metrics: - type: accuracy value: 46.986000000000004 - type: f1 value: 44.93316837240337 - task: type: Retrieval dataset: name: MTEB ArguAna type: arguana config: default split: test revision: None metrics: - type: map_at_1 value: 28.521 - type: map_at_10 value: 45.062999999999995 - type: map_at_100 value: 45.965 - type: map_at_1000 value: 45.972 - type: map_at_3 value: 40.078 - type: map_at_5 value: 43.158 - type: mrr_at_1 value: 29.232000000000003 - type: mrr_at_10 value: 45.305 - type: mrr_at_100 value: 46.213 - type: mrr_at_1000 value: 46.22 - type: mrr_at_3 value: 40.339000000000006 - type: mrr_at_5 value: 43.394 - type: ndcg_at_1 value: 28.521 - type: ndcg_at_10 value: 53.959999999999994 - type: ndcg_at_100 value: 57.691 - type: ndcg_at_1000 value: 57.858 - type: ndcg_at_3 value: 43.867 - type: ndcg_at_5 value: 49.38 - type: precision_at_1 value: 28.521 - type: precision_at_10 value: 8.222 - type: precision_at_100 value: 0.9820000000000001 - type: precision_at_1000 value: 0.1 - type: precision_at_3 value: 18.279 - type: precision_at_5 value: 13.627 - type: recall_at_1 value: 28.521 - type: recall_at_10 value: 82.219 - type: recall_at_100 value: 98.222 - type: recall_at_1000 value: 99.502 - type: recall_at_3 value: 54.836 - type: recall_at_5 value: 68.137 - task: type: Clustering dataset: name: MTEB ArxivClusteringP2P type: mteb/arxiv-clustering-p2p config: default split: test revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d metrics: - type: v_measure value: 39.409674498704625 - task: type: Reranking dataset: name: MTEB AskUbuntuDupQuestions type: mteb/askubuntudupquestions-reranking config: default split: test revision: 2000358ca161889fa9c082cb41daa8dcfb161a54 metrics: - type: map value: 61.52757354203137 - type: mrr value: 74.28241656773513 - task: type: STS dataset: name: MTEB BIOSSES type: mteb/biosses-sts config: default split: test revision: d3fb88f8f02e40887cd149695127462bbcf29b4a metrics: - type: cos_sim_pearson value: 84.39442490594014 - type: cos_sim_spearman value: 83.37599616417513 - type: euclidean_pearson value: 83.23317790460271 - type: euclidean_spearman value: 83.37599616417513 - type: manhattan_pearson value: 83.23182214744224 - type: manhattan_spearman value: 83.5428674363298 - task: type: Classification dataset: name: MTEB Banking77Classification type: mteb/banking77 config: default split: test revision: 0fd18e25b25c072e09e0d92ab615fda904d66300 metrics: - type: accuracy value: 81.93181818181819 - type: f1 value: 81.0852312152688 - task: type: Retrieval dataset: name: MTEB CQADupstackEnglishRetrieval type: BeIR/cqadupstack config: default split: test revision: None metrics: - type: map_at_1 value: 28.784 - type: map_at_10 value: 38.879000000000005 - type: map_at_100 value: 40.161 - type: map_at_1000 value: 40.291 - type: map_at_3 value: 36.104 - type: map_at_5 value: 37.671 - type: mrr_at_1 value: 35.924 - type: mrr_at_10 value: 44.471 - type: mrr_at_100 value: 45.251000000000005 - type: mrr_at_1000 value: 45.296 - type: mrr_at_3 value: 42.367 - type: mrr_at_5 value: 43.635000000000005 - type: ndcg_at_1 value: 35.924 - type: ndcg_at_10 value: 44.369 - type: ndcg_at_100 value: 48.925999999999995 - type: ndcg_at_1000 value: 50.964 - type: ndcg_at_3 value: 40.416999999999994 - type: ndcg_at_5 value: 42.309999999999995 - type: precision_at_1 value: 35.924 - type: precision_at_10 value: 8.344 - type: precision_at_100 value: 1.367 - type: precision_at_1000 value: 0.181 - type: precision_at_3 value: 19.469 - type: precision_at_5 value: 13.771 - type: recall_at_1 value: 28.784 - type: recall_at_10 value: 53.92400000000001 - type: recall_at_100 value: 72.962 - type: recall_at_1000 value: 85.90100000000001 - type: recall_at_3 value: 42.574 - type: recall_at_5 value: 47.798 - task: type: Classification dataset: name: MTEB EmotionClassification type: mteb/emotion config: default split: test revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37 metrics: - type: accuracy value: 50.16499999999999 - type: f1 value: 43.57906972116264 - task: type: Retrieval dataset: name: MTEB FiQA2018 type: fiqa config: default split: test revision: None metrics: - type: map_at_1 value: 20.737 - type: map_at_10 value: 33.566 - type: map_at_100 value: 35.367 - type: map_at_1000 value: 35.546 - type: map_at_3 value: 29.881999999999998 - type: map_at_5 value: 31.818 - type: mrr_at_1 value: 41.975 - type: mrr_at_10 value: 50.410999999999994 - type: mrr_at_100 value: 51.172 - type: mrr_at_1000 value: 51.214999999999996 - type: mrr_at_3 value: 48.611 - type: mrr_at_5 value: 49.522 - type: ndcg_at_1 value: 41.975 - type: ndcg_at_10 value: 41.299 - type: ndcg_at_100 value: 47.768 - type: ndcg_at_1000 value: 50.882000000000005 - type: ndcg_at_3 value: 38.769 - type: ndcg_at_5 value: 39.106 - type: precision_at_1 value: 41.975 - type: precision_at_10 value: 11.296000000000001 - type: precision_at_100 value: 1.7840000000000003 - type: precision_at_1000 value: 0.23500000000000001 - type: precision_at_3 value: 26.029000000000003 - type: precision_at_5 value: 18.457 - type: recall_at_1 value: 20.737 - type: recall_at_10 value: 47.284 - type: recall_at_100 value: 71.286 - type: recall_at_1000 value: 89.897 - type: recall_at_3 value: 35.411 - type: recall_at_5 value: 39.987 - task: type: Classification dataset: name: MTEB ImdbClassification type: mteb/imdb config: default split: test revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7 metrics: - type: accuracy value: 87.84 - type: ap value: 82.68294664793142 - type: f1 value: 87.8226441992267 - task: type: Classification dataset: name: MTEB MTOPDomainClassification (en) type: mteb/mtop_domain config: en split: test revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf metrics: - type: accuracy value: 93.35841313269493 - type: f1 value: 93.060022693275 - task: type: Classification dataset: name: MTEB MTOPIntentClassification (en) type: mteb/mtop_intent config: en split: test revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba metrics: - type: accuracy value: 66.58002735978113 - type: f1 value: 46.995919480823055 - task: type: Classification dataset: name: MTEB MassiveIntentClassification (en) type: mteb/amazon_massive_intent config: en split: test revision: 31efe3c427b0bae9c22cbb560b8f15491cc6bed7 metrics: - type: accuracy value: 71.07935440484196 - type: f1 value: 69.13197875645403 - task: type: Classification dataset: name: MTEB MassiveScenarioClassification (en) type: mteb/amazon_massive_scenario config: en split: test revision: 7d571f92784cd94a019292a1f45445077d0ef634 metrics: - type: accuracy value: 76.63752521856087 - type: f1 value: 75.61348469613843 - task: type: Retrieval dataset: name: MTEB NFCorpus type: nfcorpus config: default split: test revision: None metrics: - type: map_at_1 value: 5.234 - type: map_at_10 value: 11.718 - type: map_at_100 value: 14.396 - type: map_at_1000 value: 15.661 - type: map_at_3 value: 8.951 - type: map_at_5 value: 10.233 - type: mrr_at_1 value: 43.034 - type: mrr_at_10 value: 52.161 - type: mrr_at_100 value: 52.729000000000006 - type: mrr_at_1000 value: 52.776 - type: mrr_at_3 value: 50.671 - type: mrr_at_5 value: 51.476 - type: ndcg_at_1 value: 41.331 - type: ndcg_at_10 value: 31.411 - type: ndcg_at_100 value: 28.459 - type: ndcg_at_1000 value: 37.114000000000004 - type: ndcg_at_3 value: 37.761 - type: ndcg_at_5 value: 35.118 - type: precision_at_1 value: 43.034 - type: precision_at_10 value: 22.878999999999998 - type: precision_at_100 value: 7.093000000000001 - type: precision_at_1000 value: 1.9560000000000002 - type: precision_at_3 value: 35.707 - type: precision_at_5 value: 30.279 - type: recall_at_1 value: 5.234 - type: recall_at_10 value: 14.745 - type: recall_at_100 value: 28.259 - type: recall_at_1000 value: 59.16400000000001 - type: recall_at_3 value: 10.08 - type: recall_at_5 value: 11.985 - task: type: STS dataset: name: MTEB SICK-R type: mteb/sickr-sts config: default split: test revision: a6ea5a8cab320b040a23452cc28066d9beae2cee metrics: - type: cos_sim_pearson value: 83.33269306539026 - type: cos_sim_spearman value: 79.71441518631086 - type: euclidean_pearson value: 80.98109404189279 - type: euclidean_spearman value: 79.71444969096095 - type: manhattan_pearson value: 80.97223989357175 - type: manhattan_spearman value: 79.64929261210406 - task: type: STS dataset: name: MTEB STS12 type: mteb/sts12-sts config: default split: test revision: a0d554a64d88156834ff5ae9920b964011b16384 metrics: - type: cos_sim_pearson value: 83.7127498314437 - type: cos_sim_spearman value: 78.73426610516154 - type: euclidean_pearson value: 79.72827173736742 - type: euclidean_spearman value: 78.731973450314 - type: manhattan_pearson value: 79.71391822179304 - type: manhattan_spearman value: 78.69626503719782 - task: type: STS dataset: name: MTEB STS13 type: mteb/sts13-sts config: default split: test revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca metrics: - type: cos_sim_pearson value: 78.33449726355023 - type: cos_sim_spearman value: 79.59703323420547 - type: euclidean_pearson value: 79.87238808505464 - type: euclidean_spearman value: 79.59703323420547 - type: manhattan_pearson value: 79.5006260085966 - type: manhattan_spearman value: 79.21864659717262 - task: type: STS dataset: name: MTEB STS14 type: mteb/sts14-sts config: default split: test revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375 metrics: - type: cos_sim_pearson value: 79.00088445445654 - type: cos_sim_spearman value: 78.99977508575147 - type: euclidean_pearson value: 78.63222924140206 - type: euclidean_spearman value: 78.99976994069327 - type: manhattan_pearson value: 78.35504771673297 - type: manhattan_spearman value: 78.76306077740067 - task: type: STS dataset: name: MTEB STS15 type: mteb/sts15-sts config: default split: test revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3 metrics: - type: cos_sim_pearson value: 87.13160613452308 - type: cos_sim_spearman value: 87.81435104273643 - type: euclidean_pearson value: 87.22395745487297 - type: euclidean_spearman value: 87.81435041827874 - type: manhattan_pearson value: 87.17630476262896 - type: manhattan_spearman value: 87.76535338976686 - task: type: STS dataset: name: MTEB STS16 type: mteb/sts16-sts config: default split: test revision: 4d8694f8f0e0100860b497b999b3dbed754a0513 metrics: - type: cos_sim_pearson value: 83.76424652225954 - type: cos_sim_spearman value: 85.39745570134193 - type: euclidean_pearson value: 84.6971466556576 - type: euclidean_spearman value: 85.39745570134193 - type: manhattan_pearson value: 84.61210275324463 - type: manhattan_spearman value: 85.30727114432379 - task: type: STS dataset: name: MTEB STS17 (en-en) type: mteb/sts17-crosslingual-sts config: en-en split: test revision: af5e6fb845001ecf41f4c1e033ce921939a2a68d metrics: - type: cos_sim_pearson value: 86.87956530541486 - type: cos_sim_spearman value: 87.13412608536781 - type: euclidean_pearson value: 87.80084186244981 - type: euclidean_spearman value: 87.13412608536781 - type: manhattan_pearson value: 87.73101535306475 - type: manhattan_spearman value: 87.05897655963285 - task: type: STS dataset: name: MTEB STSBenchmark type: mteb/stsbenchmark-sts config: default split: test revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831 metrics: - type: cos_sim_pearson value: 83.70737517925419 - type: cos_sim_spearman value: 84.84687698325351 - type: euclidean_pearson value: 84.36525309890885 - type: euclidean_spearman value: 84.84688249844098 - type: manhattan_pearson value: 84.31171573973266 - type: manhattan_spearman value: 84.79550448196474 - task: type: Retrieval dataset: name: MTEB SciFact type: scifact config: default split: test revision: None metrics: - type: map_at_1 value: 48.178 - type: map_at_10 value: 59.24 - type: map_at_100 value: 59.902 - type: map_at_1000 value: 59.941 - type: map_at_3 value: 56.999 - type: map_at_5 value: 58.167 - type: mrr_at_1 value: 51.0 - type: mrr_at_10 value: 60.827 - type: mrr_at_100 value: 61.307 - type: mrr_at_1000 value: 61.341 - type: mrr_at_3 value: 59.0 - type: mrr_at_5 value: 60.033 - type: ndcg_at_1 value: 51.0 - type: ndcg_at_10 value: 64.366 - type: ndcg_at_100 value: 67.098 - type: ndcg_at_1000 value: 68.08 - type: ndcg_at_3 value: 60.409 - type: ndcg_at_5 value: 62.150000000000006 - type: precision_at_1 value: 51.0 - type: precision_at_10 value: 8.799999999999999 - type: precision_at_100 value: 1.027 - type: precision_at_1000 value: 0.11100000000000002 - type: precision_at_3 value: 24.444 - type: precision_at_5 value: 15.8 - type: recall_at_1 value: 48.178 - type: recall_at_10 value: 78.34400000000001 - type: recall_at_100 value: 90.36699999999999 - type: recall_at_1000 value: 98.0 - type: recall_at_3 value: 67.35 - type: recall_at_5 value: 71.989 - task: type: PairClassification dataset: name: MTEB SprintDuplicateQuestions type: mteb/sprintduplicatequestions-pairclassification config: default split: test revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46 metrics: - type: cos_sim_accuracy value: 99.87722772277228 - type: cos_sim_ap value: 97.32479581402639 - type: cos_sim_f1 value: 93.74369323915236 - type: cos_sim_precision value: 94.60285132382892 - type: cos_sim_recall value: 92.9 - type: dot_accuracy value: 99.87722772277228 - type: dot_ap value: 97.32479581402637 - type: dot_f1 value: 93.74369323915236 - type: dot_precision value: 94.60285132382892 - type: dot_recall value: 92.9 - type: euclidean_accuracy value: 99.87722772277228 - type: euclidean_ap value: 97.32479581402639 - type: euclidean_f1 value: 93.74369323915236 - type: euclidean_precision value: 94.60285132382892 - type: euclidean_recall value: 92.9 - type: manhattan_accuracy value: 99.87524752475248 - type: manhattan_ap value: 97.29133330261223 - type: manhattan_f1 value: 93.59359359359361 - type: manhattan_precision value: 93.687374749499 - type: manhattan_recall value: 93.5 - type: max_accuracy value: 99.87722772277228 - type: max_ap value: 97.32479581402639 - type: max_f1 value: 93.74369323915236 - task: type: Classification dataset: name: MTEB ToxicConversationsClassification type: mteb/toxic_conversations_50k config: default split: test revision: d7c0de2777da35d6aae2200a62c6e0e5af397c4c metrics: - type: accuracy value: 72.60060000000001 - type: ap value: 15.719924742317021 - type: f1 value: 56.30561683159878 - task: type: Classification dataset: name: MTEB TweetSentimentExtractionClassification type: mteb/tweet_sentiment_extraction config: default split: test revision: d604517c81ca91fe16a244d1248fc021f9ecee7a metrics: - type: accuracy value: 63.71250707413696 - type: f1 value: 63.54808116265952 - task: type: PairClassification dataset: name: MTEB TwitterSemEval2015 type: mteb/twittersemeval2015-pairclassification config: default split: test revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1 metrics: - type: cos_sim_accuracy value: 85.110568039578 - type: cos_sim_ap value: 70.28927714315245 - type: cos_sim_f1 value: 65.03893361488716 - type: cos_sim_precision value: 65.06469500924214 - type: cos_sim_recall value: 65.0131926121372 - type: dot_accuracy value: 85.110568039578 - type: dot_ap value: 70.28928082939848 - type: dot_f1 value: 65.03893361488716 - type: dot_precision value: 65.06469500924214 - type: dot_recall value: 65.0131926121372 - type: euclidean_accuracy value: 85.110568039578 - type: euclidean_ap value: 70.28928621260852 - type: euclidean_f1 value: 65.03893361488716 - type: euclidean_precision value: 65.06469500924214 - type: euclidean_recall value: 65.0131926121372 - type: manhattan_accuracy value: 85.02115992132086 - type: manhattan_ap value: 70.05813255171925 - type: manhattan_f1 value: 64.59658311510164 - type: manhattan_precision value: 61.24379285883188 - type: manhattan_recall value: 68.33773087071239 - type: max_accuracy value: 85.110568039578 - type: max_ap value: 70.28928621260852 - type: max_f1 value: 65.03893361488716 - task: type: PairClassification dataset: name: MTEB TwitterURLCorpus type: mteb/twitterurlcorpus-pairclassification config: default split: test revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf metrics: - type: cos_sim_accuracy value: 88.99949547871309 - type: cos_sim_ap value: 85.82819569154559 - type: cos_sim_f1 value: 78.37315338318439 - type: cos_sim_precision value: 74.46454564358494 - type: cos_sim_recall value: 82.71481367416075 - type: dot_accuracy value: 88.99949547871309 - type: dot_ap value: 85.82820043407936 - type: dot_f1 value: 78.37315338318439 - type: dot_precision value: 74.46454564358494 - type: dot_recall value: 82.71481367416075 - type: euclidean_accuracy value: 88.99949547871309 - type: euclidean_ap value: 85.82819622588083 - type: euclidean_f1 value: 78.37315338318439 - type: euclidean_precision value: 74.46454564358494 - type: euclidean_recall value: 82.71481367416075 - type: manhattan_accuracy value: 88.98009081383165 - type: manhattan_ap value: 85.77393389750326 - type: manhattan_f1 value: 78.38852097130243 - type: manhattan_precision value: 75.06341600901916 - type: manhattan_recall value: 82.0218663381583 - type: max_accuracy value: 88.99949547871309 - type: max_ap value: 85.82820043407936 - type: max_f1 value: 78.38852097130243 ---
[ "BIOSSES", "SCIFACT" ]
chenhaodev/mistral-7b-medwiki-v1
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:other", "region:us" ]
"2024-02-06T09:26:37Z"
2024-02-07T04:05:06+00:00
0
0
--- base_model: mistralai/Mistral-7B-v0.1 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: mistral-7b-medwiki-v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-medwiki-v1 This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the medical_meadow_wikidoc dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 1.0 ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Perfromance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-medwiki-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.99|± |0.0100| |professional_medicine| 0|none | 0|acc | 0.57|± |0.0498| |college_medicine | 0|none | 0|acc | 0.59|± |0.0494| |clinical_knowledge | 0|none | 0|acc | 0.58|± |0.0496| |medmcqa |Yaml |none | 0|acc | 0.40|± |0.0492| |ocn |Yaml |none | 0|acc | 0.61|± |0.0490| |aocnp |Yaml |none | 0|acc | 0.52|± |0.0502| ### Original Performance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |professional_medicine| 0|none | 0|acc | 0.64|± |0.0482| |college_medicine | 0|none | 0|acc | 0.65|± |0.0479| |clinical_knowledge | 0|none | 0|acc | 0.68|± |0.0469| |medmcqa |Yaml |none | 0|acc | 0.45|± |0.0500| |ocn |Yaml |none | 0|acc | 0.62|± |0.0488| |aocnp |Yaml |none | 0|acc | 0.47|± |0.0502|
[ "PUBMEDQA" ]
chenhaodev/mistral-7b-medqa-v1
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:other", "region:us" ]
"2024-02-07T02:28:34Z"
2024-02-07T03:05:03+00:00
0
1
--- base_model: mistralai/Mistral-7B-v0.1 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: mistral-7b-medqa-v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-medqa-v1 This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the medical_meadow_medqa dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 20 - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Performance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-medqa-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |ocn |Yaml |none | 0|acc | 0.71|± |0.0456| |professional_medicine| 0|none | 0|acc | 0.69|± |0.0465| |college_medicine | 0|none | 0|acc | 0.61|± |0.0490| |clinical_knowledge | 0|none | 0|acc | 0.63|± |0.0485| |medmcqa |Yaml |none | 0|acc | 0.41|± |0.0494| |aocnp |Yaml |none | 0|acc | 0.61|± |0.0490| ### Appendix (original performance before lora-finetune) hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |ocn |Yaml |none | 0|acc | 0.62|± |0.0488| |professional_medicine| 0|none | 0|acc | 0.64|± |0.0482| |college_medicine | 0|none | 0|acc | 0.65|± |0.0479| |clinical_knowledge | 0|none | 0|acc | 0.68|± |0.0469| |medmcqa |Yaml |none | 0|acc | 0.45|± |0.0500| |aocnp |Yaml |none | 0|acc | 0.47|± |0.0502|
[ "MEDQA", "PUBMEDQA" ]
chenhaodev/mistral-7b-mmlu-v1
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:other", "region:us" ]
"2024-02-07T05:03:57Z"
2024-02-07T05:17:54+00:00
0
0
--- base_model: mistralai/Mistral-7B-v0.1 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: mistral-7b-mmlu-v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-mmlu-v1 This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the medical_meadow_mmmlu dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 20 - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Performance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-mmlu-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |medmcqa |Yaml |none | 0|acc | 0.47|± |0.0502| |professional_medicine| 0|none | 0|acc | 0.79|± |0.0409| |college_medicine | 0|none | 0|acc | 0.72|± |0.0451| |clinical_knowledge | 0|none | 0|acc | 0.72|± |0.0451| |aocnp |Yaml |none | 0|acc | 0.56|± |0.0499| |ocn |Yaml |none | 0|acc | 0.66|± |0.0476|
[ "PUBMEDQA" ]
chenhaodev/mistral-7b-ocn-v1
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:other", "region:us" ]
"2024-02-07T06:17:58Z"
2024-02-07T06:49:53+00:00
0
0
--- base_model: mistralai/Mistral-7B-v0.1 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: mistral-7b-ocn-v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-ocn-v1 This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc_instruct dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Performance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-ocn-v1), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |medmcqa |Yaml |none | 0|acc | 0.36|± |0.0482| |professional_medicine| 0|none | 0|acc | 0.57|± |0.0498| |college_medicine | 0|none | 0|acc | 0.54|± |0.0501| |clinical_knowledge | 0|none | 0|acc | 0.62|± |0.0488| |aocnp |Yaml |none | 0|acc | 0.44|± |0.0499| |ocn |Yaml |none | 0|acc | 0.54|± |0.0501|
[ "PUBMEDQA" ]
chenhaodev/mistral-7b-ocn-v2
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:other", "region:us" ]
"2024-02-07T07:07:17Z"
2024-02-07T07:22:09+00:00
0
0
--- base_model: mistralai/Mistral-7B-v0.1 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: mistral-7b-ocn-v2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-ocn-v2 This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on the oncc_medqa_instruct dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Performance hf (pretrained=mistralai/Mistral-7B-v0.1,parallelize=True,load_in_4bit=True,peft=chenhugging/mistral-7b-ocn-v2), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.98|± |0.0141| |medmcqa |Yaml |none | 0|acc | 0.40|± |0.0492| |professional_medicine| 0|none | 0|acc | 0.69|± |0.0465| |college_medicine | 0|none | 0|acc | 0.53|± |0.0502| |clinical_knowledge | 0|none | 0|acc | 0.59|± |0.0494| |ocn |Yaml |none | 0|acc | 0.80|± |0.0402| |aocnp |Yaml |none | 0|acc | 0.63|± |0.0485|
[ "MEDQA", "PUBMEDQA" ]
chenhaodev/solar-10b-ocn-v1
chenhaodev
null
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:upstage/SOLAR-10.7B-v1.0", "base_model:adapter:upstage/SOLAR-10.7B-v1.0", "license:other", "region:us" ]
"2024-02-07T09:12:23Z"
2024-02-07T10:01:49+00:00
0
1
--- base_model: upstage/SOLAR-10.7B-v1.0 library_name: peft license: other tags: - llama-factory - lora - generated_from_trainer model-index: - name: solar-10b-ocn-v1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # solar-10b-ocn-v1 This model is a fine-tuned version of upstage/SOLAR-10.7B-v1.0 on the oncc_medqa_instruct dataset. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 1.0 - mixed_precision_training: Native AMP ### Training script CUDA_VISIBLE_DEVICES=0 python src/train_bash.py --stage sft --do_train True --model_name_or_path upstage/SOLAR-10.7B-v1.0 --template solar --finetuning_type lora --quantization_bit 4 --flash_attn True --dataset_dir data --dataset oncc_medqa_instruct --cutoff_len 1024 --learning_rate 0.0005 --num_train_epochs 1.0 --max_samples 5000 --per_device_train_batch_size 4 --gradient_accumulation_steps 4 --lr_scheduler_type cosine --max_grad_norm 1.0 --logging_steps 10 --save_steps 100 --warmup_steps 10 --neftune_noise_alpha 0.5 --lora_rank 8 --lora_dropout 0.2 --lora_target wqkv --output_dir /workspace/solar-10b-ocn-v1 --fp16 True --plot_loss True ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.1+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Performance Test script: lm_eval --model hf --model_args pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust_remote_code=True,parallelize=True,load_in_4bit=True --tasks ocn,aocnp,medmcqa,pubmedqa,mmlu_clinical_knowledge,mmlu_college_medicine,mmlu_professional_medicine --device cuda:0 --limit 100 hf (pretrained=upstage/SOLAR-10.7B-v1.0,peft=chenhugging/solar-10b-ocn-v1,trust_remote_code=True,parallelize=True,load_in_4bit=True), gen_kwargs: (None), limit: 100.0, num_fewshot: None, batch_size: 1 | Tasks |Version|Filter|n-shot| Metric |Value| |Stderr| |---------------------|-------|------|-----:|--------|----:|---|-----:| |pubmedqa | 1|none | 0|acc | 0.95|± |0.0219| |medmcqa |Yaml |none | 0|acc | 0.42|± |0.0496| |professional_medicine| 0|none | 0|acc | 0.72|± |0.0451| |college_medicine | 0|none | 0|acc | 0.67|± |0.0473| |clinical_knowledge | 0|none | 0|acc | 0.64|± |0.0482| |ocn |Yaml |none | 0|acc | 0.83|± |0.0378| |aocnp |Yaml |none | 0|acc | 0.72|± |0.0451|
[ "MEDQA", "PUBMEDQA" ]
reginaldcoghlan/qa
reginaldcoghlan
null
[ "region:us" ]
"2024-02-07T10:16:17Z"
2024-02-07T10:19:12+00:00
0
0
--- {} --- In today's digital landscape, the reliability, functionality, and performance of software are paramount to business success. At https://inoxoft.com/service/qa-consulting/, we specialize in revolutionizing your approach to testing, ensuring your products meet exemplary quality standards every step of the way. Our QA consulting services are designed to enhance efficiency, elevate user experience, and propel your business toward greater heights. As an ISO 27001 certified company and esteemed Microsoft Gold Partner, Google Cloud Partner, ISTQB Silver Partner, and recognized member of Clutch Firms that Deliver and Pangea, we bring unparalleled expertise to every project. Proud members of the Lviv IT Cluster, we are committed to setting industry standards and exceeding client expectations. Our comprehensive suite of Quality Assurance consulting services includes: Test Engineering: Our seasoned software QA consultants craft and implement robust testing frameworks tailored to your project's unique requirements. From identifying and addressing defects to verifying system performance, we cover all functional and non-functional aspects with precision. Test Management: Ensure seamless planning, execution, and delivery of QA activities throughout your project lifecycle. Our specialists align testing processes with your company goals, objectives, and quality standards, monitoring progress, and addressing issues proactively. Test Governance & Compliance: Navigating industries with stringent regulations such as healthcare, finance, and government, we define policies, procedures, and guidelines to ensure compliance. Our quality control measures mitigate risks and ensure timely addressing of compliance-related challenges. QA Audit and Improvement: We analyze your existing QA processes to identify areas for improvement, streamlining workflows, and enhancing efficiency. Leveraging automation and continuous integration practices, we optimize your testing processes for maximum efficacy. Pre-certification QA: Prepare your software products for certification and compliance with industry standards and regulations. Our comprehensive assessments, gap analyses, and mock audits ensure your solution meets the necessary criteria.
[ "CRAFT" ]
croissantllm/CroissantLLMChat-v0.1-GGML
croissantllm
text-generation
[ "legal", "code", "text-generation-inference", "art", "text-generation", "fr", "en", "dataset:croissantllm/croissant_dataset", "dataset:croissantllm/CroissantLLM-2201-sft", "dataset:cerebras/SlimPajama-627B", "dataset:uonlp/CulturaX", "dataset:pg19", "dataset:bigcode/starcoderdata", "arxiv:2402.00786", "license:mit", "region:us" ]
"2024-02-09T14:49:10Z"
2024-04-29T12:11:49+00:00
0
0
--- datasets: - croissantllm/croissant_dataset - croissantllm/CroissantLLM-2201-sft - cerebras/SlimPajama-627B - uonlp/CulturaX - pg19 - bigcode/starcoderdata language: - fr - en license: mit pipeline_tag: text-generation tags: - legal - code - text-generation-inference - art --- # CroissantLLMChat - GGML (190k steps + Chat) This model is part of the CroissantLLM initiative, and corresponds to the checkpoint after 190k steps (2.99 T) tokens and a final Chat finetuning phase. https://arxiv.org/abs/2402.00786 For best performance, it should be used with a temperature of 0.3 or more, and with the exact template described below: ```python chat = [ {"role": "user", "content": "Que puis-je faire à Marseille en hiver?"}, ] chat_input = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True) ``` corresponding to: ```python chat_input = """<|im_start|>user {USER QUERY}<|im_end|> <|im_start|>assistant\n""" ``` ## Abstract We introduce CroissantLLM, a 1.3B language model pretrained on a set of 3T English and French tokens, to bring to the research and industrial community a high-performance, fully open-sourced bilingual model that runs swiftly on consumer-grade local hardware. To that end, we pioneer the approach of training an intrinsically bilingual model with a 1:1 English-to-French pretraining data ratio, a custom tokenizer, and bilingual finetuning datasets. We release the training dataset, notably containing a French split with manually curated, high-quality, and varied data sources. To assess performance outside of English, we craft a novel benchmark, FrenchBench, consisting of an array of classification and generation tasks, covering various orthogonal aspects of model performance in the French Language. Additionally, rooted in transparency and to foster further Large Language Model research, we release codebases, and dozens of checkpoints across various model sizes, training data distributions, and training steps, as well as fine-tuned Chat models, and strong translation models. We evaluate our model through the FMTI framework, and validate 81% of the transparency criteria, far beyond the scores of even most open initiatives. This work enriches the NLP landscape, breaking away from previous English-centric work in order to strengthen our understanding of multilinguality in language models. ## Citation Our work can be cited as: ```bash @misc{faysse2024croissantllm, title={CroissantLLM: A Truly Bilingual French-English Language Model}, author={Manuel Faysse and Patrick Fernandes and Nuno M. Guerreiro and António Loison and Duarte M. Alves and Caio Corro and Nicolas Boizard and João Alves and Ricardo Rei and Pedro H. Martins and Antoni Bigata Casademunt and François Yvon and André F. T. Martins and Gautier Viaud and Céline Hudelot and Pierre Colombo}, year={2024}, eprint={2402.00786}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ## Usage This model is a Chat model, that is, it is finetuned for Chat function and works best with the provided template. #### With generate This might require a stopping criteria on <|im_end|> token. ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "croissantllm/CroissantLLMChat-v0.1" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name) generation_args = { "max_new_tokens": 256, "do_sample": True, "temperature": 0.3, "top_p": 0.90, "top_k": 40, "repetition_penalty": 1.05, "eos_token_id": [tokenizer.eos_token_id, 32000], } chat = [ {"role": "user", "content": "Qui est le président francais actuel ?"}, ] chat_input = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True) inputs = tokenizer(chat_input, return_tensors="pt").to(model.device) tokens = model.generate(**inputs, **generation_args) print(tokenizer.decode(tokens[0])) # print tokens individually print([(tokenizer.decode([tok]), tok) for tok in tokens[0].tolist()]) ``` ## Model limitations Evaluation results indicate the model is strong in its size category, and offers decent performances on writing-based tasks and internal knowledge, and very strong performance on translation tasks. The small size of the CroissantLLM model however hinders its capacity to perform more complex reasoning-based tasks, at least in a zero or few-shot manner in its generalist base or chat-model versions. This is aligned with other models of size and underlines the importance of scale for more abstract tasks. #### Knowledge Cutoff The model training dataset has a data cutoff date corresponding to the November 2023 Wikipedia dump. This is the de facto knowledge cutoff date for our base model, although a lot of information dates back further. Updated versions can be trained through continued pre-training or subsequent fine-tuning. #### Multilingual performance. CroissantLLM is mostly a French and English model. Code performance is relatively limited, and although some amount of data from other languages is included within the SlimPajama training set, out-of-the-box performance in other languages is not to be expected, although some European languages do work quite well. #### Hallucinations. CroissantLLM can hallucinate and output factually incorrect data, especially regarding complex topics. This is to be expected given the small model size, and hallucination rates seem inferior to most models of the same size category although no quantitative assessments have been conducted outside of MT-Bench experiments.
[ "CRAFT" ]
croissantllm/CroissantLLMBase-GGML
croissantllm
text-generation
[ "legal", "code", "text-generation-inference", "art", "text-generation", "fr", "en", "dataset:cerebras/SlimPajama-627B", "dataset:uonlp/CulturaX", "dataset:pg19", "dataset:bigcode/starcoderdata", "dataset:croissantllm/croissant_dataset", "arxiv:2402.00786", "license:mit", "region:us" ]
"2024-02-09T15:08:10Z"
2024-04-29T12:13:02+00:00
0
0
--- datasets: - cerebras/SlimPajama-627B - uonlp/CulturaX - pg19 - bigcode/starcoderdata - croissantllm/croissant_dataset language: - fr - en license: mit pipeline_tag: text-generation tags: - legal - code - text-generation-inference - art --- # CroissantLLM - Base GGML (190k steps, Final version) This model is part of the CroissantLLM initiative, and corresponds to the checkpoint after 190k steps (2.99 T) tokens. To play with the final model, we recommend using the Chat version: https://huggingface.co/croissantllm/CroissantLLMChat-v0.1. https://arxiv.org/abs/2402.00786 ## Abstract We introduce CroissantLLM, a 1.3B language model pretrained on a set of 3T English and French tokens, to bring to the research and industrial community a high-performance, fully open-sourced bilingual model that runs swiftly on consumer-grade local hardware. To that end, we pioneer the approach of training an intrinsically bilingual model with a 1:1 English-to-French pretraining data ratio, a custom tokenizer, and bilingual finetuning datasets. We release the training dataset, notably containing a French split with manually curated, high-quality, and varied data sources. To assess performance outside of English, we craft a novel benchmark, FrenchBench, consisting of an array of classification and generation tasks, covering various orthogonal aspects of model performance in the French Language. Additionally, rooted in transparency and to foster further Large Language Model research, we release codebases, and dozens of checkpoints across various model sizes, training data distributions, and training steps, as well as fine-tuned Chat models, and strong translation models. We evaluate our model through the FMTI framework, and validate 81% of the transparency criteria, far beyond the scores of even most open initiatives. This work enriches the NLP landscape, breaking away from previous English-centric work in order to strengthen our understanding of multilinguality in language models. ## Citation Our work can be cited as: ```bash @misc{faysse2024croissantllm, title={CroissantLLM: A Truly Bilingual French-English Language Model}, author={Manuel Faysse and Patrick Fernandes and Nuno M. Guerreiro and António Loison and Duarte M. Alves and Caio Corro and Nicolas Boizard and João Alves and Ricardo Rei and Pedro H. Martins and Antoni Bigata Casademunt and François Yvon and André F. T. Martins and Gautier Viaud and Céline Hudelot and Pierre Colombo}, year={2024}, eprint={2402.00786}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ## Usage This model is a base model, that is, it is not finetuned for Chat function and works best with few-shot prompting strategies. ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "croissantllm/CroissantLLMBase" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto") inputs = tokenizer("I am so tired I could sleep right now. -> Je suis si fatigué que je pourrais m'endormir maintenant.\nHe is heading to the market. -> Il va au marché.\nWe are running on the beach. ->", return_tensors="pt").to(model.device) tokens = model.generate(**inputs, max_length=100, do_sample=True, top_p=0.95, top_k=60, temperature=0.3) print(tokenizer.decode(tokens[0])) # remove bos token inputs = tokenizer("Capitales: France -> Paris, Italie -> Rome, Allemagne -> Berlin, Espagne ->", return_tensors="pt", add_special_tokens=True).to(model.device) tokens = model.generate(**inputs, max_length=100, do_sample=True, top_p=0.95, top_k=60) print(tokenizer.decode(tokens[0])) ```
[ "CRAFT" ]
buio/Fauno-Italian-LLM-7B
buio
null
[ "large language model", "italian large language model", "baize", "llama ", "italian", "it", "en", "dataset:andreabac3/MedQuaAD-Italian-Fauno-Baize", "dataset:andreabac3/StackOverflow-Italian-Fauno-Baize", "dataset:andreabac3/Quora-Italian-Fauno-Baize", "dataset:teelinsan/camoscio_cleaned", "license:gpl-3.0", "region:us" ]
"2024-02-13T17:18:36Z"
2024-02-13T17:24:48+00:00
0
0
--- datasets: - andreabac3/MedQuaAD-Italian-Fauno-Baize - andreabac3/StackOverflow-Italian-Fauno-Baize - andreabac3/Quora-Italian-Fauno-Baize - teelinsan/camoscio_cleaned language: - it - en license: gpl-3.0 tags: - large language model - italian large language model - baize - 'llama ' - italian --- # Fauno - Italian LLM ![image](fauno.drawio.png) Get ready to meet Fauno - the Italian language model crafted by the [RSTLess Research Group](https://rstless-lab.netlify.app/) from the Sapienza University of Rome. The talented research team behind Fauno includes [Andrea Bacciu](https://andreabac3.github.io/), [Dr. Giovanni Trappolini](https://sites.google.com/view/giovannitrappolini), [Andrea Santilli](https://www.santilli.xyz/), and [Professor Fabrizio Silvestri](https://sites.google.com/diag.uniroma1.it/fabriziosilvestri/home). Fauno represents a cutting-edge development in open-source Italian Large Language Modeling. It's trained on extensive Italian synthetic datasets, encompassing a wide range of fields such as medical data 🩺, technical content from Stack Overflow 💻, Quora discussions 💬, and Alpaca data 🦙 translated into Italian. Hence, our model is able to answer to your questions in Italian 🙋, fix your buggy code 🐛 and understand a minimum of medical literature 💊. ## The 🇮🇹 open-source version of chatGPT! Discover the capabilities of Fauno and experience the evolution of Italian language models for yourself. ![demo](screenshot_demo.png) ### Why Fauno? We started with a model called Baize, named after a legendary creature from Chinese literature. Continuing along this thematic line, we developed our Italian model based on Baize and named it Fauno, inspired by an iconic figure from Roman mythology. This choice underlines the link between the two models, while maintaining a distinctive identity rooted in Italian culture. # Did you know that you can run Fauno on Colab base? Follow this link to access a Colab notebook with our 7B version! <a target="_blank" href="https://colab.research.google.com/drive/1AepJVWS-qU910zyq-Zi7wWFQ5tthVzUe"> <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/> </a> ## 🔎 Model's details Fauno is a fine-tuned version of the LoRa weights of [Baize](https://github.com/project-baize/baize-chatbot), that is an improved version of [LLama](https://github.com/facebookresearch/llama). We translated and cleaned the data of Baize, and then we fine-tuned the 7b model using a single RTX A6000 (48GB of VRAM) with 19 hours for one epoch. - 13B: https://huggingface.co/andreabac3/Fauno-Italian-LLM-13B Fauno 30B and 65B are coming soon! ## Model initialization ```python from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig from peft import PeftModel tokenizer = LlamaTokenizer.from_pretrained("decapoda-research/llama-7b-hf") model = LlamaForCausalLM.from_pretrained( "decapoda-research/llama-7b-hf", load_in_8bit=True, device_map="auto", ) model = PeftModel.from_pretrained(model, "andreabac3/Fauno-Italian-LLM-7B") model.eval() ``` ## Inference ```python def evaluate(question: str) -> str: prompt = f"The conversation between human and AI assistant.\n[|Human|] {question}.\n[|AI|] " inputs = tokenizer(prompt, return_tensors="pt") input_ids = inputs["input_ids"].cuda() generation_output = model.generate( input_ids=input_ids, generation_config=generation_config, return_dict_in_generate=True, output_scores=True, max_new_tokens=256 ) output = tokenizer.decode(generation_output.sequences[0]).split("[|AI|]")[1] return output your_question: str = "Qual'è il significato della vita?" print(evaluate(your_question)) ``` ### Output ``` Il senso della vita è una domanda che molte persone hanno cercato di rispondere per secoli. Alla fine, il senso della vita è soggettivo e varia da persona a persona. Alcune persone credono che il senso della vita sia trovare la felicità, mentre altre credono che sia raggiungere i propri obiettivi o aiutare gli altri. Alla fine, il senso della vita è determinato dall'individuo e dai loro valori e credenze. In definitiva, il senso della vita è qualcosa che ognuno deve trovare da solo. Non c'è una risposta giusta o sbagliata, poiché ogni persona ha le proprie convinzioni e credenze. La ricerca del senso della vita può essere un viaggio lungo e difficile, ma vale la pena perseguire. ``` # 📖 Cite our work To use our translated dataset and model weights in your research, remember to cite our work. ```bibtex @misc{fauno, author = {Andrea Bacciu, Giovanni Trappolini, Andrea Santilli, Fabrizio Silvestri}, title = {Fauno: The Italian Large Language Model that will leave you senza parole!}, year = {2023}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/andreabac3/Fauno-Italian-LLM}}, } ``` ## 🔑 License This project is a derivative of Baize, and we adhere to the licensing constraints imposed by both Baize's creators and the authors of LLama. ## ⚠️ Hallucinations It is important to remark that current generation models are prone to the problem of hallucinations. So we advise you not to take their answers seriously. ## 👏 Acknowledgement - LLama - Meta AI: https://github.com/facebookresearch/llama - Baize: https://github.com/project-baize/baize-chatbot - Standford Alpaca: https://github.com/tatsu-lab/stanford_alpaca - Camoscio: https://github.com/teelinsan/camoscio #### Image Credits - llama image: https://next14.com/en/nextnews-7-march-a-new-language-model-for-meta-bing-ai-on-windows-and-the-first-tokenized-real-estate-sales/ - Fauno logo: https://www.flaticon.com/free-icon/faun_7931635?term=faun&page=1&position=1&origin=tag&related_id=7931635
[ "MEDICAL DATA" ]
GnanaPrasath/ocr_tamil
GnanaPrasath
image-to-text
[ "ocr", "optical character recognition", "text recognition", "tamil", "image-to-text", "ta", "en", "license:mit", "region:us" ]
"2024-02-14T12:38:41Z"
2024-02-14T12:49:55+00:00
0
7
--- language: - ta - en license: mit metrics: - accuracy pipeline_tag: image-to-text tags: - ocr - optical character recognition - text recognition - tamil --- <h1 align="center"> OCR Tamil - Easy, Accurate and Simple to use Tamil OCR - (ஒளி எழுத்துணரி)</h1> <p align="center">❤️️❤️️Please star✨ it if you like❤️️❤️️</p> <p align="center"> <a href="LICENSE"> <img src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/MIT.svg" alt="LICENSE"> </a> <a href="https://huggingface.co/spaces/GnanaPrasath/ocr_tamil"> <img src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/huggingface.svg" alt="HuggingSpace"> </a> <a href="https://colab.research.google.com/drive/11QPPj3EmpoIqnpuIznKeP1icxvVOjfux?usp=sharing"> <img src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/colab.svg" alt="colab"> </a> </p> <div align="center"> <p> <a href="https://github.com/gnana70/tamil_ocr"> <img width="50%" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/logo_1.gif"> </a> </p> </div> OCR Tamil can help you extract text from signboard, nameplates, storefronts etc., from Natural Scenes with high accuracy. This version of OCR is much more robust to tilted text compared to the Tesseract, Paddle OCR and Easy OCR as they are primarily built to work on the documents texts and not on natural scenes. This model is work in progress, feel free to contribute!!! ## Languages Supported 🔛 **➡️ English** **➡️ Tamil (தமிழ்)** ## Accuracy 🎯 ✔️ English > 98% ✔️ Tamil > 95% ## Comparison between Tesseract OCR and OCR Tamil ⚖️ Input Image | OCR TAMIL 🏆 | Tesseract | EasyOCR | |:--------------------------------------------------------------------------:|:--------------------:|:-----------------:|:-----------------:| | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/4.jpg"> | வாழ்கவளமுடன்✅ | க்‌ க்கஸாரகளள௮ஊகஎளமுடன்‌ ❌ | வாழக வளமுடன்❌| | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/11.jpg"> | தமிழ்வாழ்க✅ | **NO OUTPUT** ❌ | தமிழ்வாழ்க✅ | | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/2.jpg"> | கோபி ✅ | **NO OUTPUT** ❌ | ப99❌ | | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/6.jpg"> | தாம்பரம் ✅ | **NO OUTPUT** ❌ | தாம்பரம❌ | | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/1.jpg"> | நெடுஞ்சாலைத் ✅ | **NO OUTPUT** ❌ |நெடுஞ்சாலைத் ✅ | | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/5.jpg"> | அண்ணாசாலை ✅ | **NO OUTPUT** ❌ | ல@I9❌ | | <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/10.jpg"> | ரெடிமேடஸ் ❌ |**NO OUTPUT** ❌ | ரெடிமேடஸ் ❌ | **Obtained Tesseract and EasyOCR results using the [Colab notebook](https://colab.research.google.com/drive/1ylZm6afur85Pe6I10N2_tzuBFl2VIxkW?usp=sharing) with Tamil and english as language** ## How to Install and Use OCR Tamil 👨🏼‍💻 ### Quick links🌐 📔 Detailed explanation on [Medium article](https://gnana70.medium.com/ocr-tamil-easy-accurate-and-simple-to-use-tamil-ocr-b03b98697f7b). ✍️ Experiment in [Colab notebook](https://colab.research.google.com/drive/11QPPj3EmpoIqnpuIznKeP1icxvVOjfux?usp=sharing) 🤗 Test it in [Huggingface spaces](https://huggingface.co/spaces/GnanaPrasath/ocr_tamil) ### Pip install instructions🐍 In your command line, run the following command ```pip install ocr_tamil``` If you are using jupyter notebook , install like ```!pip install ocr_tamil``` ### Python Usage - Single image inference **Text Recognition only** ```python from ocr_tamil.ocr import OCR image_path = r"test_images\1.jpg" # insert your own path here ocr = OCR() text_list = ocr.predict(image_path) print(text_list[0]) ## OUTPUT : நெடுஞ்சாலைத் ``` <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/1_180.jpg"> **Text Detect + Recognition** ```python from ocr_tamil.ocr import OCR image_path = r"test_images\0.jpg" # insert your own image path here ocr = OCR(detect=True) texts = ocr.predict(image_path) print(text_list[0]) ## OUTPUT : கொடைக்கானல் Kodaikanal ``` <img width="400" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/0.jpg"> ### Batch inference mode 💻 **Text Recognition only** ```python from ocr_tamil.ocr import OCR image_path = [r"test_images\1.jpg",r"test_images\2.jpg"] # insert your own image paths here ocr = OCR() text_list = ocr.predict(image_path) for text in text_list: print(text) ## OUTPUT : நெடுஞ்சாலைத் ## OUTPUT : கோபி ``` **Text Detect + Recognition** ```python from ocr_tamil.ocr import OCR image_path = [r"test_images\0.jpg",r"test_images\tamil_sentence.jpg"] # insert your own image paths here ocr = OCR(detect=True) text_list = ocr.predict(image_path) for text in text_list: print(text) ## OUTPUT : கொடைக்கானல் Kodaikanal ## OUTPUT : செரியர் யற்கை மூலிகைகளில் இருந்து ஈர்த்தெடுக்க்கப்பட்ட வீரிய உட்பொருட்களை உள்ளடக்கி எந்த இரசாயன சேர்க்கைகளும் இல்லாமல் உருவாக்கப்பட்ட இந்தியாவின் முதல் சித்த தயாரிப்பு ``` **Tested using Python 3.10 on Windows & Linux (Ubuntu 22.04) Machines** ## Applications⚡ 1. ADAS system navigation based on the signboards + maps (hybrid approach) 🚁 2. License plate recognition 🚘 ## Limitations⛔ 1. Unable to read the text if they are present in rotated forms <p align="left"> <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/9.jpg"> <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/8.jpg"> </p> 2. Currently supports Only English and Tamil Language 3. Document Text reading capability is limited. Auto identification of Paragraph, line are not supported along with Text detection model inability to detect and crop the Tamil text leads to accuracy decrease (**WORKAROUND** Can use your own text detection model along with OCR tamil text recognition model) <p align="center"> <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/tamil_sentence.jpg"> </p> <p align="center"> <span>Cropped Text from Text detection Model</span> </p> <p align="center"> <img width="200" alt="teaser" src="https://github.com/gnana70/tamil_ocr/raw/main/test_images/tamil_sentence_crop.jpg"> </p> <p align="center"> Character **இ** missing due to text detection model error </p> **?**யற்கை மூலிகைகளில் இருந்து ஈர்த்தெடுக்கக்கப்பட்ட வீரிய உட்பொருட்களை உள்ளடக்கி எந்த இரசாயன சேர்க்கைகளும் இல்லாமல் உருவாக்கப்பட்ட இந்தியாவின் முதல் சித்த தயாரிப்பு ## Acknowledgements 👏 **Text detection** - [CRAFT TEXT DECTECTION](https://github.com/clovaai/CRAFT-pytorch) **Text recognition** - [PARSEQ](https://github.com/baudm/parseq) ```bibtex @InProceedings{bautista2022parseq, title={Scene Text Recognition with Permuted Autoregressive Sequence Models}, author={Bautista, Darwin and Atienza, Rowel}, booktitle={European Conference on Computer Vision}, pages={178--196}, month={10}, year={2022}, publisher={Springer Nature Switzerland}, address={Cham}, doi={10.1007/978-3-031-19815-1_11}, url={https://doi.org/10.1007/978-3-031-19815-1_11} } ``` ```bibtex @inproceedings{baek2019character, title={Character Region Awareness for Text Detection}, author={Baek, Youngmin and Lee, Bado and Han, Dongyoon and Yun, Sangdoo and Lee, Hwalsuk}, booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition}, pages={9365--9374}, year={2019} } ``` ## Citation ```bibtex @InProceedings{GnanaPrasath, title={Tamil OCR}, author={Gnana Prasath D}, month={01}, year={2024}, url={https://github.com/gnana70/tamil_ocr} } ``` ![logo](https://github.com/gnana70/tamil_ocr/raw/main/test_images/logo_1.gif)
[ "CRAFT" ]
ostapeno/library-phi_2-v3-10-flan-clusters-fromlucas
ostapeno
null
[ "region:us" ]
"2024-02-14T15:37:44Z"
2024-07-06T12:58:50+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | phi2_joint_lora_embed_clustersc8_2e_3epoch | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,quail_description_context_question_answer_id,quail_context_question_description_text,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer_no_instructions_,quail_context_description_question_answer_id,race_high_Taking_a_test,super_glue_multirc_1_0_2,race_middle_Select_the_best_answer,quail_context_question_description_answer_id,quail_description_context_question_answer_text,quail_context_question_answer_description_text,race_high_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_generate_span_,quail_context_question_answer_description_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_description_answer_text,quail_description_context_question_text,race_middle_Taking_a_test,quail_no_prompt_id,quail_no_prompt_text,race_middle_Select_the_best_answer_no_instructions_ | lora | | phi2_joint_lora_embed_clustersc3_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_is_the_final_step_of_the_following_process,wmt16_translate_ro_en_1_0_0,wiqa_what_might_be_the_last_step_of_the_process,wiki_bio_key_content,gem_common_gen_1_1_0,duorc_SelfRC_build_story_around_qa,app_reviews_generate_review,wiki_bio_what_content,wiki_bio_who,gem_e2e_nlg_1_1_0,cot_esnli_ii,wmt16_translate_tr_en_1_0_0,wiqa_what_is_the_missing_first_step,wiki_bio_comprehension,coqa_1_0_0,duorc_ParaphraseRC_build_story_around_qa,multi_news_1_0_0 | lora | | phi2_joint_lora_embed_clustersc0_2e_3epoch | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer,ropes_prompt_bottom_no_hint,ropes_plain_background_situation,ropes_new_situation_background_answer,ropes_given_background_situation,ropes_prompt_bottom_hint_beginning,ropes_prompt_beginning,ropes_read_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,ropes_prompt_mix,ropes_background_situation_middle | lora | | phi2_joint_lora_embed_clustersc2_2e_3epoch | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer,super_glue_record_1_0_2,wiki_hop_original_generate_object,adversarial_qa_droberta_tell_what_it_is,dbpedia_14_given_a_choice_of_categories_,wiki_hop_original_choose_best_object_affirmative_3,quac_1_0_0,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_choose_best_object_affirmative_1,adversarial_qa_dbert_answer_the_following_q,wiki_hop_original_choose_best_object_interrogative_2,adversarial_qa_droberta_question_context_answer,squad_v2_0_3_0_0,wiki_hop_original_generate_subject,wiki_bio_guess_person,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_dbert_tell_what_it_is,race_high_Write_a_multi_choice_question_options_given_,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_generate_subject_and_object,drop_2_0_0,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is | lora | | phi2_joint_lora_embed_clustersc6_2e_3epoch | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2,cot_sensemaking,super_glue_wic_1_0_2,cos_e_v1_11_rationale,anli_r3_0_1_0,dream_generate_last_utterance,paws_wiki_1_1_0,cos_e_v1_11_generate_explanation_given_text,cot_creak,stream_aqua,snli_1_1_0,cos_e_v1_11_i_think,glue_qqp_2_0_0,cos_e_v1_11_explain_why_human,anli_r2_0_1_0,anli_r1_0_1_0,glue_stsb_2_0_0,cos_e_v1_11_aligned_with_common_sense,glue_mnli_2_0_0,social_i_qa_I_was_wondering,cosmos_qa_1_0_0,glue_mrpc_2_0_0,social_i_qa_Generate_answer | lora | | phi2_joint_lora_embed_clustersc7_2e_3epoch | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question,app_reviews_convert_to_star_rating,cos_e_v1_11_question_option_description_text,social_i_qa_Show_choices_and_generate_answer,quartz_answer_question_based_on,sciq_Direct_Question_Closed_Book_,qasc_qa_with_separated_facts_3,quartz_given_the_fact_answer_the_q,quartz_answer_question_below,kilt_tasks_hotpotqa_final_exam,sciq_Multiple_Choice,wiqa_does_the_supposed_perturbation_have_an_effect,cos_e_v1_11_question_description_option_text,wiki_qa_Is_This_True_,quartz_use_info_from_question_paragraph,sciq_Direct_Question,qasc_qa_with_separated_facts_2,wiqa_which_of_the_following_is_the_supposed_perturbation,app_reviews_convert_to_rating,cos_e_v1_11_question_option_description_id,wiqa_effect_with_string_answer,qasc_qa_with_separated_facts_5,dream_baseline,quartz_having_read_above_passage,cos_e_v1_11_question_description_option_id,qasc_qa_with_separated_facts_1,cos_e_v1_11_description_question_option_text,qasc_qa_with_combined_facts_1,qasc_is_correct_1,cos_e_v1_11_description_question_option_id,social_i_qa_Check_if_a_random_answer_is_valid_or_not,sciq_Multiple_Choice_Closed_Book_,quartz_use_info_from_paragraph_question,qasc_is_correct_2,qasc_qa_with_separated_facts_4,quartz_read_passage_below_choose,quartz_paragraph_question_plain_concat,sciq_Multiple_Choice_Question_First | lora | | phi2_joint_lora_embed_clustersc1_2e_3epoch | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0,adversarial_qa_droberta_generate_question,true_case,stream_qed,huggingface_xsum,cot_esnli,cot_gsm8k,trec_1_0_0,yelp_polarity_reviews_0_2_0,lambada_1_0_0,glue_cola_2_0_0,ag_news_subset_1_0_0,gem_dart_1_1_0,math_dataset_algebra__linear_1d_1_0_0,cnn_dailymail_3_4_0,wiki_hop_original_explain_relation,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,gem_wiki_lingua_english_en_1_1_0,fix_punct,imdb_reviews_plain_text_1_0_0,race_middle_Write_a_multi_choice_question_for_the_following_article,gigaword_1_2_0,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,gem_web_nlg_en_1_1_0,word_segment,race_high_Write_a_multi_choice_question_for_the_following_article,wmt16_translate_de_en_1_0_0,cot_ecqa,aeslc_1_0_0,dream_generate_first_utterance,wmt16_translate_fi_en_1_0_0,dream_answer_to_dialogue,para_crawl_enes,adversarial_qa_dbert_generate_question,race_middle_Write_a_multi_choice_question_options_given_,wmt14_translate_fr_en_1_0_0 | lora | | phi2_joint_lora_embed_clustersc9_2e_3epoch | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0,web_questions_whats_the_answer,web_questions_question_answer,dbpedia_14_pick_one_category_for_the_following_text,kilt_tasks_hotpotqa_combining_facts,web_questions_short_general_knowledge_q,kilt_tasks_hotpotqa_straighforward_qa,adversarial_qa_dbidaf_generate_question,adversarial_qa_droberta_based_on,web_questions_get_the_answer,kilt_tasks_hotpotqa_complex_question,web_questions_potential_correct_answer,trivia_qa_rc_1_1_0,kilt_tasks_hotpotqa_formulate,adversarial_qa_dbert_based_on,adversarial_qa_dbidaf_based_on,squad_v1_1_3_0_0 | lora | | phi2_joint_lora_embed_clustersc4_2e_3epoch | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google,app_reviews_categorize_rating_using_review,race_middle_Is_this_the_right_answer,super_glue_cb_1_0_2,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Direct_Answer_to_Question,super_glue_wsc_fixed_1_0_2,cot_gsm8k_ii,unified_qa_science_inst,race_high_Is_this_the_right_answer,cot_strategyqa,cot_ecqa_ii,quarel_do_not_use,wiki_qa_exercise,wiki_qa_automatic_system,cot_creak_ii,quarel_heres_a_story,quarel_choose_between,stream_qed_ii,wiki_qa_Topic_Prediction_Question_Only,glue_qnli_2_0_0,cot_sensemaking_ii,super_glue_copa_1_0_2,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_Show_choices_and_generate_index,quarel_testing_students,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_Decide_good_answer,wiki_qa_Jeopardy_style,wiki_qa_Generate_Question_from_Topic,definite_pronoun_resolution_1_1_0,wiqa_effect_with_label_answer,glue_wnli_2_0_0,cot_qasc,cot_strategyqa_ii,quarel_logic_test,stream_aqua_ii | lora | | phi2_joint_lora_embed_clustersc5_2e_3epoch | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer,duorc_SelfRC_generate_question_by_answer,quoref_Find_Answer,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_answer_question,quoref_Found_Context_Online,quoref_Read_And_Extract_,duorc_ParaphraseRC_title_generation,duorc_ParaphraseRC_decide_worth_it,quoref_What_Is_The_Answer,duorc_ParaphraseRC_generate_question,quoref_Guess_Title_For_Context,quoref_Answer_Test,duorc_SelfRC_question_answering,duorc_SelfRC_title_generation,duorc_ParaphraseRC_generate_question_by_answer,duorc_ParaphraseRC_extract_answer,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_ParaphraseRC_question_answering,quoref_Answer_Question_Given_Context,duorc_SelfRC_extract_answer,quoref_Guess_Answer,quoref_Answer_Friend_Question,duorc_SelfRC_movie_director,duorc_SelfRC_generate_question,quoref_Given_Context_Answer_Question | lora | Last updated on: 2024-02-14 16:05:40+00:00 Cloned from pclucas14/library-phi_2-v3-10-flan-clusters
[ "SCIQ" ]
ostapeno/library-phi_2-v3_fromsordonia
ostapeno
null
[ "region:us" ]
"2024-02-14T17:40:34Z"
2024-06-22T15:33:42+00:00
0
0
--- {} --- Number of experts present in the library: 263 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | glue_sst2_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | adversarial_qa_droberta_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | adversarial_qa_dbidaf_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | app_reviews_convert_to_star_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | race_high_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | super_glue_rte_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | true_case | phi-2 | sordonia/flan-10k-flat/true_case | lora | | wiqa_what_might_be_the_first_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | quail_description_context_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | quail_context_question_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | stream_qed | phi-2 | sordonia/flan-10k-flat/stream_qed | lora | | huggingface_xsum | phi-2 | sordonia/flan-10k-flat/huggingface_xsum | lora | | cos_e_v1_11_question_option_description_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | wiqa_what_is_the_final_step_of_the_following_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | ropes_background_new_situation_answer | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | wiki_qa_found_on_google | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | cot_esnli | phi-2 | sordonia/flan-10k-flat/cot_esnli | lora | | social_i_qa_Show_choices_and_generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | cot_gsm8k | phi-2 | sordonia/flan-10k-flat/cot_gsm8k | lora | | app_reviews_categorize_rating_using_review | phi-2 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | cot_sensemaking | phi-2 | sordonia/flan-10k-flat/cot_sensemaking | lora | | trec_1_0_0 | phi-2 | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_wic_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_prompt_bottom_no_hint | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | quartz_answer_question_based_on | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | super_glue_record_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | yelp_polarity_reviews_0_2_0 | phi-2 | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | race_middle_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | quoref_Context_Contains_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | cos_e_v1_11_rationale | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | natural_questions_open_1_0_0 | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | ropes_plain_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | web_questions_whats_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | anli_r3_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | duorc_SelfRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | quoref_Find_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | duorc_ParaphraseRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | sciq_Direct_Question_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | qasc_qa_with_separated_facts_3 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | lambada_1_0_0 | phi-2 | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | quartz_given_the_fact_answer_the_q | phi-2 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | super_glue_cb_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | quartz_answer_question_below | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | duorc_ParaphraseRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | wmt16_translate_ro_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | dream_generate_last_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | wiki_qa_Topic_Prediction_Answer_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | kilt_tasks_hotpotqa_final_exam | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | glue_cola_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | race_high_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quail_context_description_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | ag_news_subset_1_0_0 | phi-2 | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | paws_wiki_1_1_0 | phi-2 | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | sciq_Multiple_Choice | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | wiki_qa_Direct_Answer_to_Question | phi-2 | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | gem_dart_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | cos_e_v1_11_generate_explanation_given_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | wiki_hop_original_generate_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | race_high_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | wiqa_what_might_be_the_last_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | wiki_bio_key_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | quoref_Found_Context_Online | phi-2 | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | super_glue_wsc_fixed_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | phi-2 | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | adversarial_qa_droberta_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | cos_e_v1_11_question_description_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | gem_common_gen_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | quoref_Read_And_Extract_ | phi-2 | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | cot_creak | phi-2 | sordonia/flan-10k-flat/cot_creak | lora | | cot_gsm8k_ii | phi-2 | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | duorc_ParaphraseRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | hellaswag_1_1_0 | phi-2 | sordonia/flan-10k-flat/hellaswag_1_1_0 | lora | | wiki_qa_Is_This_True_ | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | math_dataset_algebra__linear_1d_1_0_0 | phi-2 | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | unified_qa_science_inst | phi-2 | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quartz_use_info_from_question_paragraph | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | web_questions_question_answer | phi-2 | sordonia/flan-10k-flat/web_questions_question_answer | lora | | duorc_ParaphraseRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | stream_aqua | phi-2 | sordonia/flan-10k-flat/stream_aqua | lora | | dbpedia_14_pick_one_category_for_the_following_text | phi-2 | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | super_glue_multirc_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | dbpedia_14_given_a_choice_of_categories_ | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | sciq_Direct_Question | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | kilt_tasks_hotpotqa_combining_facts | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quoref_What_Is_The_Answer | phi-2 | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | web_questions_short_general_knowledge_q | phi-2 | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | qasc_qa_with_separated_facts_2 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | phi-2 | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | cnn_dailymail_3_4_0 | phi-2 | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | duorc_ParaphraseRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | race_middle_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | kilt_tasks_hotpotqa_straighforward_qa | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | duorc_SelfRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | adversarial_qa_dbidaf_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | snli_1_1_0 | phi-2 | sordonia/flan-10k-flat/snli_1_1_0 | lora | | app_reviews_convert_to_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiki_hop_original_choose_best_object_affirmative_3 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | quail_context_question_description_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | cos_e_v1_11_i_think | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | quoref_Guess_Title_For_Context | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | quac_1_0_0 | phi-2 | sordonia/flan-10k-flat/quac_1_0_0 | lora | | cos_e_v1_11_question_option_description_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | quoref_Answer_Test | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | wiki_hop_original_choose_best_object_interrogative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | duorc_SelfRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | wiki_hop_original_explain_relation | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | ropes_new_situation_background_answer | phi-2 | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | race_high_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | quail_description_context_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | cot_strategyqa | phi-2 | sordonia/flan-10k-flat/cot_strategyqa | lora | | ropes_given_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | quail_context_question_answer_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | cot_ecqa_ii | phi-2 | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | ropes_prompt_bottom_hint_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | gem_wiki_lingua_english_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | glue_qqp_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | ai2_arc_ARC_Challenge_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Challenge_1_0_0 | lora | | fix_punct | phi-2 | sordonia/flan-10k-flat/fix_punct | lora | | wiqa_effect_with_string_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | adversarial_qa_droberta_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | imdb_reviews_plain_text_1_0_0 | phi-2 | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | race_high_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | race_middle_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | quarel_do_not_use | phi-2 | sordonia/flan-10k-flat/quarel_do_not_use | lora | | duorc_SelfRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | qasc_qa_with_separated_facts_5 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | wiki_qa_exercise | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | duorc_ParaphraseRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | web_questions_get_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | wiki_hop_original_choose_best_object_affirmative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | duorc_ParaphraseRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | dream_baseline | phi-2 | sordonia/flan-10k-flat/dream_baseline | lora | | adversarial_qa_dbert_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | gigaword_1_2_0 | phi-2 | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | ropes_prompt_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | quail_context_question_answer_description_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | duorc_SelfRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | kilt_tasks_hotpotqa_complex_question | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | quartz_having_read_above_passage | phi-2 | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | quail_context_description_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | cos_e_v1_11_question_description_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | ropes_read_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | gem_web_nlg_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | adversarial_qa_droberta_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | qasc_qa_with_separated_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | wiki_qa_automatic_system | phi-2 | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_plain_bottom_hint | phi-2 | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | duorc_SelfRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | duorc_ParaphraseRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | cos_e_v1_11_explain_why_human | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | word_segment | phi-2 | sordonia/flan-10k-flat/word_segment | lora | | cot_creak_ii | phi-2 | sordonia/flan-10k-flat/cot_creak_ii | lora | | anli_r2_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | cos_e_v1_11_description_question_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | quarel_heres_a_story | phi-2 | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | qasc_qa_with_combined_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | app_reviews_generate_review | phi-2 | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_bio_what_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | qasc_is_correct_1 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | quoref_Answer_Question_Given_Context | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | squad_v2_0_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | web_questions_potential_correct_answer | phi-2 | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | trivia_qa_rc_1_1_0 | phi-2 | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | wmt16_translate_de_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | cos_e_v1_11_description_question_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_hop_original_generate_subject | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | ropes_plain_no_background | phi-2 | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | quarel_choose_between | phi-2 | sordonia/flan-10k-flat/quarel_choose_between | lora | | stream_qed_ii | phi-2 | sordonia/flan-10k-flat/stream_qed_ii | lora | | wiki_bio_guess_person | phi-2 | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | anli_r1_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | quail_context_description_question_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | cot_ecqa | phi-2 | sordonia/flan-10k-flat/cot_ecqa | lora | | quail_context_question_description_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | wiki_bio_who | phi-2 | sordonia/flan-10k-flat/wiki_bio_who | lora | | wiki_qa_Topic_Prediction_Question_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | glue_stsb_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | cos_e_v1_11_aligned_with_common_sense | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | aeslc_1_0_0 | phi-2 | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | dream_generate_first_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | wmt16_translate_fi_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | adversarial_qa_dbidaf_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | dream_answer_to_dialogue | phi-2 | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | glue_qnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | adversarial_qa_droberta_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | cot_sensemaking_ii | phi-2 | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | adversarial_qa_dbert_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | glue_mnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | quail_description_context_question_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | super_glue_copa_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | phi-2 | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | social_i_qa_Generate_the_question_from_the_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | social_i_qa_Show_choices_and_generate_index | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | kilt_tasks_hotpotqa_formulate | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | gem_e2e_nlg_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | para_crawl_enes | phi-2 | sordonia/flan-10k-flat/para_crawl_enes | lora | | ai2_arc_ARC_Easy_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Easy_1_0_0 | lora | | duorc_SelfRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | sciq_Multiple_Choice_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | race_high_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | race_middle_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | social_i_qa_I_was_wondering | phi-2 | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | adversarial_qa_dbert_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | quoref_Guess_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | race_middle_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | quartz_use_info_from_paragraph_question | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | quoref_Answer_Friend_Question | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | qasc_is_correct_2 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wmt14_translate_fr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_testing_students | phi-2 | sordonia/flan-10k-flat/quarel_testing_students | lora | | piqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/piqa_1_0_0 | lora | | wiki_hop_original_choose_best_object_affirmative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | qasc_qa_with_separated_facts_4 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | duorc_SelfRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | cosmos_qa_1_0_0 | phi-2 | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | cot_esnli_ii | phi-2 | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quail_no_prompt_id | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | wmt16_translate_tr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | wiki_qa_Decide_good_answer | phi-2 | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | wiki_qa_Jeopardy_style | phi-2 | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | adversarial_qa_dbert_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | winogrande_1_1_0 | phi-2 | sordonia/flan-10k-flat/winogrande_1_1_0 | lora | | bool_q_1_0_0 | phi-2 | sordonia/flan-10k-flat/bool_q_1_0_0 | lora | | duorc_SelfRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | wiki_qa_Generate_Question_from_Topic | phi-2 | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | wiki_hop_original_generate_subject_and_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | adversarial_qa_dbidaf_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiqa_what_is_the_missing_first_step | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | quartz_read_passage_below_choose | phi-2 | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | definite_pronoun_resolution_1_1_0 | phi-2 | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | quail_no_prompt_text | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | wiqa_effect_with_label_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | drop_2_0_0 | phi-2 | sordonia/flan-10k-flat/drop_2_0_0 | lora | | race_middle_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | glue_wnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | wiki_bio_comprehension | phi-2 | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | glue_mrpc_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cot_qasc | phi-2 | sordonia/flan-10k-flat/cot_qasc | lora | | adversarial_qa_dbert_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | openbookqa_0_1_0 | phi-2 | sordonia/flan-10k-flat/openbookqa_0_1_0 | lora | | quoref_Given_Context_Answer_Question | phi-2 | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | coqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | quartz_paragraph_question_plain_concat | phi-2 | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | adversarial_qa_dbidaf_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | ropes_prompt_mix | phi-2 | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | social_i_qa_Generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | cot_strategyqa_ii | phi-2 | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | quarel_logic_test | phi-2 | sordonia/flan-10k-flat/quarel_logic_test | lora | | duorc_ParaphraseRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | stream_aqua_ii | phi-2 | sordonia/flan-10k-flat/stream_aqua_ii | lora | | multi_news_1_0_0 | phi-2 | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | ropes_background_situation_middle | phi-2 | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | sciq_Multiple_Choice_Question_First | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | squad_v1_1_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | Last updated on: 2024-02-14 17:41:44+00:00 Clone from sordonia/library-phi_2-v3
[ "SCIQ" ]
tmohoric-ewc/safer-skin
tmohoric-ewc
tabular-regression
[ "sklearn", "skops", "tabular-regression", "license:mit", "region:us" ]
"2024-02-15T23:19:43Z"
2024-02-15T23:21:29+00:00
0
0
--- library_name: sklearn license: mit tags: - sklearn - skops - tabular-regression model_format: pickle model_file: MLR-model.pkl widget: - structuredData: CAS: - 696-71-9 - 94-02-0 - 15128-82-2 CID: - 12766.0 - 7170.0 - 27057.0 CanonicalSMILES: - canonical: OC1CCCCCCC1 original: C1CCCC(CCC1)O - canonical: CCOC(=O)CC(=O)c1ccccc1 original: CCOC(=O)CC(=O)C1=CC=CC=C1 - canonical: O=[N+]([O-])c1ncccc1O original: C1=CC(=C(N=C1)[N+](=O)[O-])O Cor1-C420 Adduct (M+H): - no Adduct - no Adduct - no Adduct Cor1-C420 Depletion 24 h (%): - 1.0 - 1.0 - 1.0 Cor1-C420 Dimer (%): - 2.0 - 5.0 - 4.0 Cor1-C420 Kmax (1/mM/min): - 6.979399898264935e-06 - 6.979399898264935e-06 - 6.979399898264935e-06 DPRA Cysteine depletion (%): - .nan - 11.2 - .nan DPRA Lysine depletion (%): - .nan - 0.9 - .nan InChI: - InChI=1S/C8H16O/c9-8-6-4-2-1-3-5-7-8/h8-9H,1-7H2 - InChI=1S/C11H12O3/c1-2-14-11(13)8-10(12)9-6-4-3-5-7-9/h3-7H,2,8H2,1H3 - InChI=1S/C5H4N2O3/c8-4-2-1-3-6-5(4)7(9)10/h1-3,8H InChIKey: - FHADSMKORVFYOS-UHFFFAOYSA-N - GKKZMYDNDDMXSE-UHFFFAOYSA-N - QBPDSKPWYWIHGA-UHFFFAOYSA-N IsomericSMILES: - canonical: OC1CCCCCCC1 original: C1CCCC(CCC1)O - canonical: CCOC(=O)CC(=O)c1ccccc1 original: CCOC(=O)CC(=O)C1=CC=CC=C1 - canonical: O=[N+]([O-])c1ncccc1O original: C1=CC(=C(N=C1)[N+](=O)[O-])O KeratinoSens EC1.5 (uM): - 249.6822169 - 62.9764329 - 4000.0 KeratinoSens EC3 (uM): - 4000.0 - 689.0 - 4000.0 KeratinoSens IC50 (uM): - 4000.0 - 4000.0 - 4000.0 KeratinoSens Imax: - 2.830997136 - 3.299878249 - 1.036847118 KeratinoSens Log EC1.5 (uM): - 2.3973876117256947 - 1.7991780577657597 - 3.6020599913279625 KeratinoSens Log IC50 (uM): - 3.6020599913279625 - 3.6020599913279625 - 3.6020599913279625 LLNA EC3 (%): - 100.0 - 100.0 - 100.0 LLNA Log EC3 (%): - 2.0 - 2.0 - 2.0 MW: - 128.21 - 192.21 - 140.1 OPERA Boiling point (°C): - 186.863 - 276.068 - 323.069 OPERA Henry constant (atm/m3): - 7.84426e-06 - 5.86618e-07 - 9.47507e-08 OPERA Log D at pH 5.5: - 2.36 - 1.87 - -0.01 OPERA Log D at pH 7.4: - 2.36 - 1.87 - -1.69 OPERA Melting point (°C): - 25.1423 - 49.3271 - 128.292 OPERA Octanol-air partition coefficient Log Koa: - 6.08747 - 6.56126 - 6.36287 OPERA Octanol-water partition coefficient LogP: - 2.3597 - 1.86704 - 0.398541 OPERA Vapour pressure (mm Hg): - 0.0839894 - 0.000406705 - 0.00472604 OPERA Water solubility (mol/L): - 0.0510404 - 0.01476 - 0.0416421 OPERA pKaa: - 10.68 - .nan - 5.31 OPERA pKab: - .nan - .nan - .nan SMILES: - canonical: OC1CCCCCCC1 original: OC1CCCCCCC1 - canonical: CCOC(=O)CC(=O)c1ccccc1 original: CCOC(=O)CC(=O)c1ccccc1 - canonical: O=[N+]([O-])c1ncccc1O original: OC1=CC=CN=C1[N+]([O-])=O TIMES Log Vapour pressure (Pa): - 0.8564932564458658 - -0.2851674875666674 - -0.9385475209128068 Vapour pressure (Pa): - 7.1861 - 0.5186 - 0.1152 cLogP: - 2.285000000003492 - 1.206000000005588 - 0.5590000000020154 hCLAT CV75 (ug/mL): - .nan - 571.0951916 - .nan hCLAT Call: - .nan - 0.0 - .nan hCLAT EC150 (ug/mL): - .nan - .nan - .nan hCLAT EC200 (ug/mL): - .nan - .nan - .nan hCLAT MIT (ug/mL): - .nan - .nan - .nan kDPRA Call: [] kDPRA Log rate (1/s/M): - .nan - .nan - .nan --- # Model description [More Information Needed] ## Intended uses & limitations [More Information Needed] ## Training Procedure [More Information Needed] ### Hyperparameters <details> <summary> Click to expand </summary> | Hyperparameter | Value | |------------------|---------| | copy_X | True | | fit_intercept | True | | n_jobs | | | positive | False | </details> ### Model Plot <style>#sk-container-id-1 {color: black;background-color: white;}#sk-container-id-1 pre{padding: 0;}#sk-container-id-1 div.sk-toggleable {background-color: white;}#sk-container-id-1 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-container-id-1 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-container-id-1 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-container-id-1 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-container-id-1 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-container-id-1 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-container-id-1 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-container-id-1 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-container-id-1 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-container-id-1 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-container-id-1 div.sk-estimator:hover {background-color: #d4ebff;}#sk-container-id-1 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-container-id-1 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: 0;}#sk-container-id-1 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;position: relative;}#sk-container-id-1 div.sk-item {position: relative;z-index: 1;}#sk-container-id-1 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;position: relative;}#sk-container-id-1 div.sk-item::before, #sk-container-id-1 div.sk-parallel-item::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: -1;}#sk-container-id-1 div.sk-parallel-item {display: flex;flex-direction: column;z-index: 1;position: relative;background-color: white;}#sk-container-id-1 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-container-id-1 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-container-id-1 div.sk-parallel-item:only-child::after {width: 0;}#sk-container-id-1 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;}#sk-container-id-1 div.sk-label label {font-family: monospace;font-weight: bold;display: inline-block;line-height: 1.2em;}#sk-container-id-1 div.sk-label-container {text-align: center;}#sk-container-id-1 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-container-id-1 div.sk-text-repr-fallback {display: none;}</style><div id="sk-container-id-1" class="sk-top-container" style="overflow: auto;"><div class="sk-text-repr-fallback"><pre>LinearRegression()</pre><b>In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook. <br />On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.</b></div><div class="sk-container" hidden><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-1" type="checkbox" checked><label for="sk-estimator-id-1" class="sk-toggleable__label sk-toggleable__label-arrow">LinearRegression</label><div class="sk-toggleable__content"><pre>LinearRegression()</pre></div></div></div></div></div> ## Evaluation Results [More Information Needed] # How to Get Started with the Model [More Information Needed] # Model Card Authors This model card is written by following authors: [More Information Needed] # Model Card Contact You can contact the model card authors through following channels: [More Information Needed] # Citation Below you can find information related to citation. **BibTeX:** ``` [More Information Needed] ``` # model_card_authors Tomaz Mohoric # limitations This model is intended for educational purposes. # model_description This is a multiple linear regression model on a skin sensitisation dataset.
[ "CAS" ]
tacotaco122/as
tacotaco122
null
[ "region:us" ]
"2024-02-19T01:50:57Z"
2024-02-19T01:51:16+00:00
0
0
--- {} --- Hello, here to inquire about a position at your business, hope your week has been fantastic. Working with customers is fulfilling aside from the most basic response. I've worked at a lot of warehouses and I enjoy working with people more than anything, tired of being stuck in a warehouse. People are pretty uplifting, even the ones I may not see eye to with. Still, always the few out the day that show a nice gesture and have good intentions for the world, some are hilarious. At least one person out of the week does positivity towards you, or you witness good gestures, humor is subjective. Is a great skill or trait you learn the more people you talk to is an amazing trait to learn and craft.
[ "CRAFT" ]
yy0514/llama2-7b-chat-qlora-lek-train-for-medqa-2-epochs
yy0514
null
[ "safetensors", "generated_from_trainer", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:finetune:meta-llama/Llama-2-7b-chat-hf", "region:us" ]
"2024-02-19T13:28:38Z"
2024-02-19T14:30:51+00:00
0
0
--- base_model: meta-llama/Llama-2-7b-chat-hf tags: - generated_from_trainer model-index: - name: llama2-7b-chat-qlora-lek-train-for-medqa-2-epochs results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama2-7b-chat-qlora-lek-train-for-medqa-2-epochs This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - num_epochs: 2 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.1 - Tokenizers 0.15.2
[ "MEDQA" ]
BlueSkyThompson/Opi
BlueSkyThompson
null
[ "license:unknown", "region:us" ]
"2024-02-20T07:16:57Z"
2024-02-20T07:38:50+00:00
0
1
--- license: unknown --- hello. Opi is my first projet please bear with me as i am learning to write python as well as the essentials to LLMs. i Currenttly suffer from a range of mental illness and am a student game designer. my Goals for Opi and all following are build a model that is to build something i can build and reflect with my interests are in music and philosophy as well as esports and game theory.
[ "BEAR" ]
DrishtiSharma/llama-7b-chat-hf-medqa-packing-true-padding-left
DrishtiSharma
null
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:NousResearch/Llama-2-7b-chat-hf", "base_model:adapter:NousResearch/Llama-2-7b-chat-hf", "region:us" ]
"2024-02-21T11:07:45Z"
2024-02-21T11:08:04+00:00
0
0
--- base_model: NousResearch/Llama-2-7b-chat-hf library_name: peft tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-chat-hf-medqa-packing-true-padding-left results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-chat-hf-medqa-packing-true-padding-left This model is a fine-tuned version of [NousResearch/Llama-2-7b-chat-hf](https://huggingface.co/NousResearch/Llama-2-7b-chat-hf) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.01 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.39.0.dev0 - Pytorch 2.1.0+cu121 - Datasets 2.17.2.dev0 - Tokenizers 0.15.2
[ "MEDQA" ]
RoversX/GPT_SOVITS_LeiJun_V1
RoversX
null
[ "region:us" ]
"2024-02-24T02:47:10Z"
2024-10-07T12:05:09+00:00
0
0
--- {} --- **Disclaimer** This speech model and the content it generates are for personal learning and research purposes only. We do not guarantee the copyright of the content generated by this model, and users should bear the relevant risks during use. Please note that it is strictly prohibited to use this model for commercial purposes or any illegal purposes. By using this model, you agree to abide by the above terms and bear all consequences that may arise from the use of this model. We are not responsible for any loss or damage caused by the use of this model.
[ "BEAR" ]
vyshnavids03/my-pet-dog
vyshnavids03
text-to-image
[ "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "region:us" ]
"2024-02-28T12:05:01Z"
2024-02-28T12:06:37+00:00
0
0
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### My-Pet-Dog Dreambooth model trained by vyshnavids03 following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 4MK21CS059 Sample pictures of this concept: ![0](https://huggingface.co/vyshnavids03/my-pet-dog/resolve/main/sample_images/teddy-bear--on-the-grass-.jpeg)
[ "BEAR" ]
vyshnavids03/my-pet-dog-xzg
vyshnavids03
text-to-image
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
"2024-02-28T12:21:14Z"
2024-02-28T12:25:21+00:00
0
0
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### My-Pet-Dog-XZG Dreambooth model trained by vyshnavids03 following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 4MK21CS059 Sample pictures of this concept: ![0](https://huggingface.co/vyshnavids03/my-pet-dog-xzg/resolve/main/sample_images/teddy-bear--on-the-grass-.jpeg)
[ "BEAR" ]
frizai/Pulse-v1
frizai
null
[ "multi-modal", "all-in-one", "chatbot", "gpt-4", "gpt-3.5 turbo", "dall-e", "whisper", "meta-llama", "image", "3d", "audio", "en", "doi:10.57967/hf/2222", "license:apache-2.0", "region:us" ]
"2024-03-02T09:03:33Z"
2024-03-02T09:10:05+00:00
0
1
--- language: - en license: apache-2.0 tags: - multi-modal - all-in-one - chatbot - gpt-4 - gpt-3.5 turbo - dall-e - whisper - meta-llama - image - 3d - audio --- # Pulse AI PulseAI, an innovative product by FrizAI, stands at the forefront of auto-generative AI technology. Leveraging the power of quantum computing and advanced machine learning techniques, PulseAI specializes in creating diverse forms of digital content. From generating intricate text compositions to developing sophisticated web applications, PulseAI taps into the immense potential of image-based prompts to revolutionize content creation. ## Overview Pulse-AI is a dynamic Flask-based web application that integrates OpenAI's cutting-edge models. Aimed at delivering a seamless user experience, Pulse-AI provides an intuitive interface for both novices and experts to explore the capabilities of AI in image and text generation, as well as code development. ## Core Features - **Image Generation**: Utilize OpenAI's state-of-the-art Image API to transform ideas into vivid visual representations. - **Text Generation**: Harness the power of advanced language models for creating compelling and contextually relevant textual content. - **Code Generation**: Leverage AI to generate efficient and effective code snippets, enhancing productivity in software development. # Installation and Setup Embark on your journey with Pulse-AI by following these setup instructions: 1. **Clone the Repository** ```bash git clone https://github.com/Will-Langhart/Pulse-AI.git cd Pulse-AI ``` 2. **Initialize the Virtual Environment** ```bash python3 -m venv venv source venv/bin/activate ``` 3. **Dependency Installation** ```bash pip install -r requirements.txt ``` 4. **Configure Environment Variables** Create a `.env` file at the project root with your OpenAI API key: ```makefile OPENAI_API_KEY=<your_api_key_here> ``` 5. **Launch the Application** ```bash flask run ``` ## Virtual Environment ```bash deactivate # If you're currently in the virtual environment rm -rf venv python3 -m venv venv source venv/bin/activate pip install -r requirements.txt ``` # Package Downolads Dotnev ```bash pip install python-dotenv ``` OpenAI ```bash pip install openai ``` Flask-SQLAlchemy ```bash pip install flask-sqlalchemy ``` Verify Installation ```bash pip freeze ``` # GitHub Repository Changes and Updates 1. Fetch GitHub Repository Status ```bash git status ``` 2. Fetch the latest changes ```bash git fetch origin ``` 3. Merge changes into local branch ```bash git merge origin/main ``` 4. Merge head force ```bash git push origin HEAD:main ``` ## Usage Instructions Access the world of AI-powered content creation by navigating to `http://127.0.0.1:5000`. The platform offers: - **Image Generation**: Input your creative prompts and watch as AI brings them to life. - **Textual Content Creation**: Explore AI's ability to craft narratives, articles, and more. - **Code Synthesis**: Generate code snippets to accelerate your development process. ## Contribution Guidelines Join the Pulse-AI community and contribute to the evolution of AI-driven content creation: 1. Fork the repository and create a new feature branch. 2. Make your contributions and commit them with clear, descriptive messages. 3. Push your changes and initiate a pull request for review. ## License Pulse-AI is under the MIT License. For detailed information, refer to the `LICENSE` file. ## Contact and Further Information - **Contact**: [Your Name](mailto:[email protected]) - **Twitter**: [@YourTwitter](https://twitter.com/YourTwitter) - **Project Link**: [PulseAI on GitHub](https://github.com/Will-Langhart/Pulse-AI) --- PulseAI is a testament to FrizAI's commitment to advancing AI technology, making it accessible and transformative for various industries and creative endeavors.
[ "CRAFT" ]
flashingtt/imagenet-r
flashingtt
null
[ "en", "arxiv:1409.0575", "arxiv:1912.07726", "arxiv:1811.12231", "arxiv:2109.13228", "license:other", "region:us" ]
"2024-03-04T13:44:10Z"
2024-03-05T07:10:31+00:00
0
0
--- language: - en license: - other annotations_creators: - crowdsourced language_creators: - crowdsourced license_details: imagenet-agreement multilinguality: - monolingual paperswithcode_id: imagenet-1k-1 pretty_name: ImageNet size_categories: - 1M<n<10M source_datasets: - original task_categories: - image-classification task_ids: - multi-class-image-classification extra_gated_prompt: 'By clicking on “Access repository” below, you also agree to ImageNet Terms of Access: [RESEARCHER_FULLNAME] (the "Researcher") has requested permission to use the ImageNet database (the "Database") at Princeton University and Stanford University. In exchange for such permission, Researcher hereby agrees to the following terms and conditions: 1. Researcher shall use the Database only for non-commercial research and educational purposes. 2. Princeton University, Stanford University and Hugging Face make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. 3. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the ImageNet team, Princeton University, Stanford University and Hugging Face, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher''s use of the Database, including but not limited to Researcher''s use of any copies of copyrighted images that he or she may create from the Database. 4. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. 5. Princeton University, Stanford University and Hugging Face reserve the right to terminate Researcher''s access to the Database at any time. 6. If Researcher is employed by a for-profit, commercial entity, Researcher''s employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer. 7. The law of the State of New Jersey shall apply to all disputes under this agreement.' dataset_info: features: - name: image dtype: image - name: label dtype: class_label: names: 0: tench, Tinca tinca 1: goldfish, Carassius auratus 2: great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias 3: tiger shark, Galeocerdo cuvieri 4: hammerhead, hammerhead shark 5: electric ray, crampfish, numbfish, torpedo 6: stingray 7: cock 8: hen 9: ostrich, Struthio camelus 10: brambling, Fringilla montifringilla 11: goldfinch, Carduelis carduelis 12: house finch, linnet, Carpodacus mexicanus 13: junco, snowbird 14: indigo bunting, indigo finch, indigo bird, Passerina cyanea 15: robin, American robin, Turdus migratorius 16: bulbul 17: jay 18: magpie 19: chickadee 20: water ouzel, dipper 21: kite 22: bald eagle, American eagle, Haliaeetus leucocephalus 23: vulture 24: great grey owl, great gray owl, Strix nebulosa 25: European fire salamander, Salamandra salamandra 26: common newt, Triturus vulgaris 27: eft 28: spotted salamander, Ambystoma maculatum 29: axolotl, mud puppy, Ambystoma mexicanum 30: bullfrog, Rana catesbeiana 31: tree frog, tree-frog 32: tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui 33: loggerhead, loggerhead turtle, Caretta caretta 34: leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea 35: mud turtle 36: terrapin 37: box turtle, box tortoise 38: banded gecko 39: common iguana, iguana, Iguana iguana 40: American chameleon, anole, Anolis carolinensis 41: whiptail, whiptail lizard 42: agama 43: frilled lizard, Chlamydosaurus kingi 44: alligator lizard 45: Gila monster, Heloderma suspectum 46: green lizard, Lacerta viridis 47: African chameleon, Chamaeleo chamaeleon 48: Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis 49: African crocodile, Nile crocodile, Crocodylus niloticus 50: American alligator, Alligator mississipiensis 51: triceratops 52: thunder snake, worm snake, Carphophis amoenus 53: ringneck snake, ring-necked snake, ring snake 54: hognose snake, puff adder, sand viper 55: green snake, grass snake 56: king snake, kingsnake 57: garter snake, grass snake 58: water snake 59: vine snake 60: night snake, Hypsiglena torquata 61: boa constrictor, Constrictor constrictor 62: rock python, rock snake, Python sebae 63: Indian cobra, Naja naja 64: green mamba 65: sea snake 66: horned viper, cerastes, sand viper, horned asp, Cerastes cornutus 67: diamondback, diamondback rattlesnake, Crotalus adamanteus 68: sidewinder, horned rattlesnake, Crotalus cerastes 69: trilobite 70: harvestman, daddy longlegs, Phalangium opilio 71: scorpion 72: black and gold garden spider, Argiope aurantia 73: barn spider, Araneus cavaticus 74: garden spider, Aranea diademata 75: black widow, Latrodectus mactans 76: tarantula 77: wolf spider, hunting spider 78: tick 79: centipede 80: black grouse 81: ptarmigan 82: ruffed grouse, partridge, Bonasa umbellus 83: prairie chicken, prairie grouse, prairie fowl 84: peacock 85: quail 86: partridge 87: African grey, African gray, Psittacus erithacus 88: macaw 89: sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita 90: lorikeet 91: coucal 92: bee eater 93: hornbill 94: hummingbird 95: jacamar 96: toucan 97: drake 98: red-breasted merganser, Mergus serrator 99: goose 100: black swan, Cygnus atratus 101: tusker 102: echidna, spiny anteater, anteater 103: platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus 104: wallaby, brush kangaroo 105: koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus 106: wombat 107: jellyfish 108: sea anemone, anemone 109: brain coral 110: flatworm, platyhelminth 111: nematode, nematode worm, roundworm 112: conch 113: snail 114: slug 115: sea slug, nudibranch 116: chiton, coat-of-mail shell, sea cradle, polyplacophore 117: chambered nautilus, pearly nautilus, nautilus 118: Dungeness crab, Cancer magister 119: rock crab, Cancer irroratus 120: fiddler crab 121: king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica 122: American lobster, Northern lobster, Maine lobster, Homarus americanus 123: spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish 124: crayfish, crawfish, crawdad, crawdaddy 125: hermit crab 126: isopod 127: white stork, Ciconia ciconia 128: black stork, Ciconia nigra 129: spoonbill 130: flamingo 131: little blue heron, Egretta caerulea 132: American egret, great white heron, Egretta albus 133: bittern 134: crane 135: limpkin, Aramus pictus 136: European gallinule, Porphyrio porphyrio 137: American coot, marsh hen, mud hen, water hen, Fulica americana 138: bustard 139: ruddy turnstone, Arenaria interpres 140: red-backed sandpiper, dunlin, Erolia alpina 141: redshank, Tringa totanus 142: dowitcher 143: oystercatcher, oyster catcher 144: pelican 145: king penguin, Aptenodytes patagonica 146: albatross, mollymawk 147: grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus 148: killer whale, killer, orca, grampus, sea wolf, Orcinus orca 149: dugong, Dugong dugon 150: sea lion 151: Chihuahua 152: Japanese spaniel 153: Maltese dog, Maltese terrier, Maltese 154: Pekinese, Pekingese, Peke 155: Shih-Tzu 156: Blenheim spaniel 157: papillon 158: toy terrier 159: Rhodesian ridgeback 160: Afghan hound, Afghan 161: basset, basset hound 162: beagle 163: bloodhound, sleuthhound 164: bluetick 165: black-and-tan coonhound 166: Walker hound, Walker foxhound 167: English foxhound 168: redbone 169: borzoi, Russian wolfhound 170: Irish wolfhound 171: Italian greyhound 172: whippet 173: Ibizan hound, Ibizan Podenco 174: Norwegian elkhound, elkhound 175: otterhound, otter hound 176: Saluki, gazelle hound 177: Scottish deerhound, deerhound 178: Weimaraner 179: Staffordshire bullterrier, Staffordshire bull terrier 180: American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier 181: Bedlington terrier 182: Border terrier 183: Kerry blue terrier 184: Irish terrier 185: Norfolk terrier 186: Norwich terrier 187: Yorkshire terrier 188: wire-haired fox terrier 189: Lakeland terrier 190: Sealyham terrier, Sealyham 191: Airedale, Airedale terrier 192: cairn, cairn terrier 193: Australian terrier 194: Dandie Dinmont, Dandie Dinmont terrier 195: Boston bull, Boston terrier 196: miniature schnauzer 197: giant schnauzer 198: standard schnauzer 199: Scotch terrier, Scottish terrier, Scottie 200: Tibetan terrier, chrysanthemum dog 201: silky terrier, Sydney silky 202: soft-coated wheaten terrier 203: West Highland white terrier 204: Lhasa, Lhasa apso 205: flat-coated retriever 206: curly-coated retriever 207: golden retriever 208: Labrador retriever 209: Chesapeake Bay retriever 210: German short-haired pointer 211: vizsla, Hungarian pointer 212: English setter 213: Irish setter, red setter 214: Gordon setter 215: Brittany spaniel 216: clumber, clumber spaniel 217: English springer, English springer spaniel 218: Welsh springer spaniel 219: cocker spaniel, English cocker spaniel, cocker 220: Sussex spaniel 221: Irish water spaniel 222: kuvasz 223: schipperke 224: groenendael 225: malinois 226: briard 227: kelpie 228: komondor 229: Old English sheepdog, bobtail 230: Shetland sheepdog, Shetland sheep dog, Shetland 231: collie 232: Border collie 233: Bouvier des Flandres, Bouviers des Flandres 234: Rottweiler 235: German shepherd, German shepherd dog, German police dog, alsatian 236: Doberman, Doberman pinscher 237: miniature pinscher 238: Greater Swiss Mountain dog 239: Bernese mountain dog 240: Appenzeller 241: EntleBucher 242: boxer 243: bull mastiff 244: Tibetan mastiff 245: French bulldog 246: Great Dane 247: Saint Bernard, St Bernard 248: Eskimo dog, husky 249: malamute, malemute, Alaskan malamute 250: Siberian husky 251: dalmatian, coach dog, carriage dog 252: affenpinscher, monkey pinscher, monkey dog 253: basenji 254: pug, pug-dog 255: Leonberg 256: Newfoundland, Newfoundland dog 257: Great Pyrenees 258: Samoyed, Samoyede 259: Pomeranian 260: chow, chow chow 261: keeshond 262: Brabancon griffon 263: Pembroke, Pembroke Welsh corgi 264: Cardigan, Cardigan Welsh corgi 265: toy poodle 266: miniature poodle 267: standard poodle 268: Mexican hairless 269: timber wolf, grey wolf, gray wolf, Canis lupus 270: white wolf, Arctic wolf, Canis lupus tundrarum 271: red wolf, maned wolf, Canis rufus, Canis niger 272: coyote, prairie wolf, brush wolf, Canis latrans 273: dingo, warrigal, warragal, Canis dingo 274: dhole, Cuon alpinus 275: African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus 276: hyena, hyaena 277: red fox, Vulpes vulpes 278: kit fox, Vulpes macrotis 279: Arctic fox, white fox, Alopex lagopus 280: grey fox, gray fox, Urocyon cinereoargenteus 281: tabby, tabby cat 282: tiger cat 283: Persian cat 284: Siamese cat, Siamese 285: Egyptian cat 286: cougar, puma, catamount, mountain lion, painter, panther, Felis concolor 287: lynx, catamount 288: leopard, Panthera pardus 289: snow leopard, ounce, Panthera uncia 290: jaguar, panther, Panthera onca, Felis onca 291: lion, king of beasts, Panthera leo 292: tiger, Panthera tigris 293: cheetah, chetah, Acinonyx jubatus 294: brown bear, bruin, Ursus arctos 295: American black bear, black bear, Ursus americanus, Euarctos americanus 296: ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus 297: sloth bear, Melursus ursinus, Ursus ursinus 298: mongoose 299: meerkat, mierkat 300: tiger beetle 301: ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle 302: ground beetle, carabid beetle 303: long-horned beetle, longicorn, longicorn beetle 304: leaf beetle, chrysomelid 305: dung beetle 306: rhinoceros beetle 307: weevil 308: fly 309: bee 310: ant, emmet, pismire 311: grasshopper, hopper 312: cricket 313: walking stick, walkingstick, stick insect 314: cockroach, roach 315: mantis, mantid 316: cicada, cicala 317: leafhopper 318: lacewing, lacewing fly 319: dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk 320: damselfly 321: admiral 322: ringlet, ringlet butterfly 323: monarch, monarch butterfly, milkweed butterfly, Danaus plexippus 324: cabbage butterfly 325: sulphur butterfly, sulfur butterfly 326: lycaenid, lycaenid butterfly 327: starfish, sea star 328: sea urchin 329: sea cucumber, holothurian 330: wood rabbit, cottontail, cottontail rabbit 331: hare 332: Angora, Angora rabbit 333: hamster 334: porcupine, hedgehog 335: fox squirrel, eastern fox squirrel, Sciurus niger 336: marmot 337: beaver 338: guinea pig, Cavia cobaya 339: sorrel 340: zebra 341: hog, pig, grunter, squealer, Sus scrofa 342: wild boar, boar, Sus scrofa 343: warthog 344: hippopotamus, hippo, river horse, Hippopotamus amphibius 345: ox 346: water buffalo, water ox, Asiatic buffalo, Bubalus bubalis 347: bison 348: ram, tup 349: bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis 350: ibex, Capra ibex 351: hartebeest 352: impala, Aepyceros melampus 353: gazelle 354: Arabian camel, dromedary, Camelus dromedarius 355: llama 356: weasel 357: mink 358: polecat, fitch, foulmart, foumart, Mustela putorius 359: black-footed ferret, ferret, Mustela nigripes 360: otter 361: skunk, polecat, wood pussy 362: badger 363: armadillo 364: three-toed sloth, ai, Bradypus tridactylus 365: orangutan, orang, orangutang, Pongo pygmaeus 366: gorilla, Gorilla gorilla 367: chimpanzee, chimp, Pan troglodytes 368: gibbon, Hylobates lar 369: siamang, Hylobates syndactylus, Symphalangus syndactylus 370: guenon, guenon monkey 371: patas, hussar monkey, Erythrocebus patas 372: baboon 373: macaque 374: langur 375: colobus, colobus monkey 376: proboscis monkey, Nasalis larvatus 377: marmoset 378: capuchin, ringtail, Cebus capucinus 379: howler monkey, howler 380: titi, titi monkey 381: spider monkey, Ateles geoffroyi 382: squirrel monkey, Saimiri sciureus 383: Madagascar cat, ring-tailed lemur, Lemur catta 384: indri, indris, Indri indri, Indri brevicaudatus 385: Indian elephant, Elephas maximus 386: African elephant, Loxodonta africana 387: lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens 388: giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca 389: barracouta, snoek 390: eel 391: coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch 392: rock beauty, Holocanthus tricolor 393: anemone fish 394: sturgeon 395: gar, garfish, garpike, billfish, Lepisosteus osseus 396: lionfish 397: puffer, pufferfish, blowfish, globefish 398: abacus 399: abaya 400: academic gown, academic robe, judge's robe 401: accordion, piano accordion, squeeze box 402: acoustic guitar 403: aircraft carrier, carrier, flattop, attack aircraft carrier 404: airliner 405: airship, dirigible 406: altar 407: ambulance 408: amphibian, amphibious vehicle 409: analog clock 410: apiary, bee house 411: apron 412: ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin 413: assault rifle, assault gun 414: backpack, back pack, knapsack, packsack, rucksack, haversack 415: bakery, bakeshop, bakehouse 416: balance beam, beam 417: balloon 418: ballpoint, ballpoint pen, ballpen, Biro 419: Band Aid 420: banjo 421: bannister, banister, balustrade, balusters, handrail 422: barbell 423: barber chair 424: barbershop 425: barn 426: barometer 427: barrel, cask 428: barrow, garden cart, lawn cart, wheelbarrow 429: baseball 430: basketball 431: bassinet 432: bassoon 433: bathing cap, swimming cap 434: bath towel 435: bathtub, bathing tub, bath, tub 436: beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon 437: beacon, lighthouse, beacon light, pharos 438: beaker 439: bearskin, busby, shako 440: beer bottle 441: beer glass 442: bell cote, bell cot 443: bib 444: bicycle-built-for-two, tandem bicycle, tandem 445: bikini, two-piece 446: binder, ring-binder 447: binoculars, field glasses, opera glasses 448: birdhouse 449: boathouse 450: bobsled, bobsleigh, bob 451: bolo tie, bolo, bola tie, bola 452: bonnet, poke bonnet 453: bookcase 454: bookshop, bookstore, bookstall 455: bottlecap 456: bow 457: bow tie, bow-tie, bowtie 458: brass, memorial tablet, plaque 459: brassiere, bra, bandeau 460: breakwater, groin, groyne, mole, bulwark, seawall, jetty 461: breastplate, aegis, egis 462: broom 463: bucket, pail 464: buckle 465: bulletproof vest 466: bullet train, bullet 467: butcher shop, meat market 468: cab, hack, taxi, taxicab 469: caldron, cauldron 470: candle, taper, wax light 471: cannon 472: canoe 473: can opener, tin opener 474: cardigan 475: car mirror 476: carousel, carrousel, merry-go-round, roundabout, whirligig 477: carpenter's kit, tool kit 478: carton 479: car wheel 480: cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM 481: cassette 482: cassette player 483: castle 484: catamaran 485: CD player 486: cello, violoncello 487: cellular telephone, cellular phone, cellphone, cell, mobile phone 488: chain 489: chainlink fence 490: chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour 491: chain saw, chainsaw 492: chest 493: chiffonier, commode 494: chime, bell, gong 495: china cabinet, china closet 496: Christmas stocking 497: church, church building 498: cinema, movie theater, movie theatre, movie house, picture palace 499: cleaver, meat cleaver, chopper 500: cliff dwelling 501: cloak 502: clog, geta, patten, sabot 503: cocktail shaker 504: coffee mug 505: coffeepot 506: coil, spiral, volute, whorl, helix 507: combination lock 508: computer keyboard, keypad 509: confectionery, confectionary, candy store 510: container ship, containership, container vessel 511: convertible 512: corkscrew, bottle screw 513: cornet, horn, trumpet, trump 514: cowboy boot 515: cowboy hat, ten-gallon hat 516: cradle 517: crane2 518: crash helmet 519: crate 520: crib, cot 521: Crock Pot 522: croquet ball 523: crutch 524: cuirass 525: dam, dike, dyke 526: desk 527: desktop computer 528: dial telephone, dial phone 529: diaper, nappy, napkin 530: digital clock 531: digital watch 532: dining table, board 533: dishrag, dishcloth 534: dishwasher, dish washer, dishwashing machine 535: disk brake, disc brake 536: dock, dockage, docking facility 537: dogsled, dog sled, dog sleigh 538: dome 539: doormat, welcome mat 540: drilling platform, offshore rig 541: drum, membranophone, tympan 542: drumstick 543: dumbbell 544: Dutch oven 545: electric fan, blower 546: electric guitar 547: electric locomotive 548: entertainment center 549: envelope 550: espresso maker 551: face powder 552: feather boa, boa 553: file, file cabinet, filing cabinet 554: fireboat 555: fire engine, fire truck 556: fire screen, fireguard 557: flagpole, flagstaff 558: flute, transverse flute 559: folding chair 560: football helmet 561: forklift 562: fountain 563: fountain pen 564: four-poster 565: freight car 566: French horn, horn 567: frying pan, frypan, skillet 568: fur coat 569: garbage truck, dustcart 570: gasmask, respirator, gas helmet 571: gas pump, gasoline pump, petrol pump, island dispenser 572: goblet 573: go-kart 574: golf ball 575: golfcart, golf cart 576: gondola 577: gong, tam-tam 578: gown 579: grand piano, grand 580: greenhouse, nursery, glasshouse 581: grille, radiator grille 582: grocery store, grocery, food market, market 583: guillotine 584: hair slide 585: hair spray 586: half track 587: hammer 588: hamper 589: hand blower, blow dryer, blow drier, hair dryer, hair drier 590: hand-held computer, hand-held microcomputer 591: handkerchief, hankie, hanky, hankey 592: hard disc, hard disk, fixed disk 593: harmonica, mouth organ, harp, mouth harp 594: harp 595: harvester, reaper 596: hatchet 597: holster 598: home theater, home theatre 599: honeycomb 600: hook, claw 601: hoopskirt, crinoline 602: horizontal bar, high bar 603: horse cart, horse-cart 604: hourglass 605: iPod 606: iron, smoothing iron 607: jack-o'-lantern 608: jean, blue jean, denim 609: jeep, landrover 610: jersey, T-shirt, tee shirt 611: jigsaw puzzle 612: jinrikisha, ricksha, rickshaw 613: joystick 614: kimono 615: knee pad 616: knot 617: lab coat, laboratory coat 618: ladle 619: lampshade, lamp shade 620: laptop, laptop computer 621: lawn mower, mower 622: lens cap, lens cover 623: letter opener, paper knife, paperknife 624: library 625: lifeboat 626: lighter, light, igniter, ignitor 627: limousine, limo 628: liner, ocean liner 629: lipstick, lip rouge 630: Loafer 631: lotion 632: loudspeaker, speaker, speaker unit, loudspeaker system, speaker system 633: loupe, jeweler's loupe 634: lumbermill, sawmill 635: magnetic compass 636: mailbag, postbag 637: mailbox, letter box 638: maillot 639: maillot, tank suit 640: manhole cover 641: maraca 642: marimba, xylophone 643: mask 644: matchstick 645: maypole 646: maze, labyrinth 647: measuring cup 648: medicine chest, medicine cabinet 649: megalith, megalithic structure 650: microphone, mike 651: microwave, microwave oven 652: military uniform 653: milk can 654: minibus 655: miniskirt, mini 656: minivan 657: missile 658: mitten 659: mixing bowl 660: mobile home, manufactured home 661: Model T 662: modem 663: monastery 664: monitor 665: moped 666: mortar 667: mortarboard 668: mosque 669: mosquito net 670: motor scooter, scooter 671: mountain bike, all-terrain bike, off-roader 672: mountain tent 673: mouse, computer mouse 674: mousetrap 675: moving van 676: muzzle 677: nail 678: neck brace 679: necklace 680: nipple 681: notebook, notebook computer 682: obelisk 683: oboe, hautboy, hautbois 684: ocarina, sweet potato 685: odometer, hodometer, mileometer, milometer 686: oil filter 687: organ, pipe organ 688: oscilloscope, scope, cathode-ray oscilloscope, CRO 689: overskirt 690: oxcart 691: oxygen mask 692: packet 693: paddle, boat paddle 694: paddlewheel, paddle wheel 695: padlock 696: paintbrush 697: pajama, pyjama, pj's, jammies 698: palace 699: panpipe, pandean pipe, syrinx 700: paper towel 701: parachute, chute 702: parallel bars, bars 703: park bench 704: parking meter 705: passenger car, coach, carriage 706: patio, terrace 707: pay-phone, pay-station 708: pedestal, plinth, footstall 709: pencil box, pencil case 710: pencil sharpener 711: perfume, essence 712: Petri dish 713: photocopier 714: pick, plectrum, plectron 715: pickelhaube 716: picket fence, paling 717: pickup, pickup truck 718: pier 719: piggy bank, penny bank 720: pill bottle 721: pillow 722: ping-pong ball 723: pinwheel 724: pirate, pirate ship 725: pitcher, ewer 726: plane, carpenter's plane, woodworking plane 727: planetarium 728: plastic bag 729: plate rack 730: plow, plough 731: plunger, plumber's helper 732: Polaroid camera, Polaroid Land camera 733: pole 734: police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria 735: poncho 736: pool table, billiard table, snooker table 737: pop bottle, soda bottle 738: pot, flowerpot 739: potter's wheel 740: power drill 741: prayer rug, prayer mat 742: printer 743: prison, prison house 744: projectile, missile 745: projector 746: puck, hockey puck 747: punching bag, punch bag, punching ball, punchball 748: purse 749: quill, quill pen 750: quilt, comforter, comfort, puff 751: racer, race car, racing car 752: racket, racquet 753: radiator 754: radio, wireless 755: radio telescope, radio reflector 756: rain barrel 757: recreational vehicle, RV, R.V. 758: reel 759: reflex camera 760: refrigerator, icebox 761: remote control, remote 762: restaurant, eating house, eating place, eatery 763: revolver, six-gun, six-shooter 764: rifle 765: rocking chair, rocker 766: rotisserie 767: rubber eraser, rubber, pencil eraser 768: rugby ball 769: rule, ruler 770: running shoe 771: safe 772: safety pin 773: saltshaker, salt shaker 774: sandal 775: sarong 776: sax, saxophone 777: scabbard 778: scale, weighing machine 779: school bus 780: schooner 781: scoreboard 782: screen, CRT screen 783: screw 784: screwdriver 785: seat belt, seatbelt 786: sewing machine 787: shield, buckler 788: shoe shop, shoe-shop, shoe store 789: shoji 790: shopping basket 791: shopping cart 792: shovel 793: shower cap 794: shower curtain 795: ski 796: ski mask 797: sleeping bag 798: slide rule, slipstick 799: sliding door 800: slot, one-armed bandit 801: snorkel 802: snowmobile 803: snowplow, snowplough 804: soap dispenser 805: soccer ball 806: sock 807: solar dish, solar collector, solar furnace 808: sombrero 809: soup bowl 810: space bar 811: space heater 812: space shuttle 813: spatula 814: speedboat 815: spider web, spider's web 816: spindle 817: sports car, sport car 818: spotlight, spot 819: stage 820: steam locomotive 821: steel arch bridge 822: steel drum 823: stethoscope 824: stole 825: stone wall 826: stopwatch, stop watch 827: stove 828: strainer 829: streetcar, tram, tramcar, trolley, trolley car 830: stretcher 831: studio couch, day bed 832: stupa, tope 833: submarine, pigboat, sub, U-boat 834: suit, suit of clothes 835: sundial 836: sunglass 837: sunglasses, dark glasses, shades 838: sunscreen, sunblock, sun blocker 839: suspension bridge 840: swab, swob, mop 841: sweatshirt 842: swimming trunks, bathing trunks 843: swing 844: switch, electric switch, electrical switch 845: syringe 846: table lamp 847: tank, army tank, armored combat vehicle, armoured combat vehicle 848: tape player 849: teapot 850: teddy, teddy bear 851: television, television system 852: tennis ball 853: thatch, thatched roof 854: theater curtain, theatre curtain 855: thimble 856: thresher, thrasher, threshing machine 857: throne 858: tile roof 859: toaster 860: tobacco shop, tobacconist shop, tobacconist 861: toilet seat 862: torch 863: totem pole 864: tow truck, tow car, wrecker 865: toyshop 866: tractor 867: trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi 868: tray 869: trench coat 870: tricycle, trike, velocipede 871: trimaran 872: tripod 873: triumphal arch 874: trolleybus, trolley coach, trackless trolley 875: trombone 876: tub, vat 877: turnstile 878: typewriter keyboard 879: umbrella 880: unicycle, monocycle 881: upright, upright piano 882: vacuum, vacuum cleaner 883: vase 884: vault 885: velvet 886: vending machine 887: vestment 888: viaduct 889: violin, fiddle 890: volleyball 891: waffle iron 892: wall clock 893: wallet, billfold, notecase, pocketbook 894: wardrobe, closet, press 895: warplane, military plane 896: washbasin, handbasin, washbowl, lavabo, wash-hand basin 897: washer, automatic washer, washing machine 898: water bottle 899: water jug 900: water tower 901: whiskey jug 902: whistle 903: wig 904: window screen 905: window shade 906: Windsor tie 907: wine bottle 908: wing 909: wok 910: wooden spoon 911: wool, woolen, woollen 912: worm fence, snake fence, snake-rail fence, Virginia fence 913: wreck 914: yawl 915: yurt 916: web site, website, internet site, site 917: comic book 918: crossword puzzle, crossword 919: street sign 920: traffic light, traffic signal, stoplight 921: book jacket, dust cover, dust jacket, dust wrapper 922: menu 923: plate 924: guacamole 925: consomme 926: hot pot, hotpot 927: trifle 928: ice cream, icecream 929: ice lolly, lolly, lollipop, popsicle 930: French loaf 931: bagel, beigel 932: pretzel 933: cheeseburger 934: hotdog, hot dog, red hot 935: mashed potato 936: head cabbage 937: broccoli 938: cauliflower 939: zucchini, courgette 940: spaghetti squash 941: acorn squash 942: butternut squash 943: cucumber, cuke 944: artichoke, globe artichoke 945: bell pepper 946: cardoon 947: mushroom 948: Granny Smith 949: strawberry 950: orange 951: lemon 952: fig 953: pineapple, ananas 954: banana 955: jackfruit, jak, jack 956: custard apple 957: pomegranate 958: hay 959: carbonara 960: chocolate sauce, chocolate syrup 961: dough 962: meat loaf, meatloaf 963: pizza, pizza pie 964: potpie 965: burrito 966: red wine 967: espresso 968: cup 969: eggnog 970: alp 971: bubble 972: cliff, drop, drop-off 973: coral reef 974: geyser 975: lakeside, lakeshore 976: promontory, headland, head, foreland 977: sandbar, sand bar 978: seashore, coast, seacoast, sea-coast 979: valley, vale 980: volcano 981: ballplayer, baseball player 982: groom, bridegroom 983: scuba diver 984: rapeseed 985: daisy 986: yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum 987: corn 988: acorn 989: hip, rose hip, rosehip 990: buckeye, horse chestnut, conker 991: coral fungus 992: agaric 993: gyromitra 994: stinkhorn, carrion fungus 995: earthstar 996: hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa 997: bolete 998: ear, spike, capitulum 999: toilet tissue, toilet paper, bathroom tissue splits: - name: test num_bytes: 13613661561 num_examples: 100000 - name: train num_bytes: 146956944242 num_examples: 1281167 - name: validation num_bytes: 6709003386 num_examples: 50000 download_size: 166009941208 dataset_size: 167279609189 --- # Dataset Card for ImageNet ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://image-net.org/index.php - **Repository:** - **Paper:** https://arxiv.org/abs/1409.0575 - **Leaderboard:** https://paperswithcode.com/sota/image-classification-on-imagenet?tag_filter=171 - **Point of Contact:** mailto: [email protected] ### Dataset Summary ILSVRC 2012, commonly known as 'ImageNet' is an image dataset organized according to the WordNet hierarchy. Each meaningful concept in WordNet, possibly described by multiple words or word phrases, is called a "synonym set" or "synset". There are more than 100,000 synsets in WordNet, majority of them are nouns (80,000+). ImageNet aims to provide on average 1000 images to illustrate each synset. Images of each concept are quality-controlled and human-annotated. 💡 This dataset provides access to ImageNet (ILSVRC) 2012 which is the most commonly used **subset** of ImageNet. This dataset spans 1000 object classes and contains 1,281,167 training images, 50,000 validation images and 100,000 test images. The version also has the [patch](https://drive.google.com/file/d/16RYnHpVOW0XKCsn3G3S9GTHUyoV2-4WX/view) which fixes some of the corrupted test set images already applied. For full ImageNet dataset presented in [[2]](https://ieeexplore.ieee.org/abstract/document/5206848), please check the download section of the [main website](https://image-net.org/download-images.php). ### Supported Tasks and Leaderboards - `image-classification`: The goal of this task is to classify a given image into one of 1000 ImageNet classes. The leaderboard is available [here](https://paperswithcode.com/sota/image-classification-on-imagenet?tag_filter=171). To evaluate the `imagenet-classification` accuracy on the test split, one must first create an account at https://image-net.org. This account must be approved by the site administrator. After the account is created, one can submit the results to the test server at https://image-net.org/challenges/LSVRC/eval_server.php The submission consists of several ASCII text files corresponding to multiple tasks. The task of interest is "Classification submission (top-5 cls error)". A sample of an exported text file looks like the following: ``` 670 778 794 387 650 217 691 564 909 364 737 369 430 531 124 755 930 755 512 152 ``` The export format is described in full in "readme.txt" within the 2013 development kit available here: https://image-net.org/data/ILSVRC/2013/ILSVRC2013_devkit.tgz. Please see the section entitled "3.3 CLS-LOC submission format". Briefly, the format of the text file is 100,000 lines corresponding to each image in the test split. Each line of integers correspond to the rank-ordered, top 5 predictions for each test image. The integers are 1-indexed corresponding to the line number in the corresponding labels file. See `imagenet2012_labels.txt`. ### Languages The class labels in the dataset are in English. ## Dataset Structure ### Data Instances An example looks like below: ``` { 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=384x512 at 0x276021C5EB8>, 'label': 23 } ``` ### Data Fields The data instances have the following fields: - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `label`: an `int` classification label. -1 for `test` set as the labels are missing. The labels are indexed based on a sorted list of synset ids such as `n07565083` which we automatically map to original class names. The original dataset is divided into folders based on these synset ids. To get a mapping from original synset names, use the file [LOC_synset_mapping.txt](https://www.kaggle.com/competitions/imagenet-object-localization-challenge/data?select=LOC_synset_mapping.txt) available on Kaggle challenge page. You can also use `dataset_instance.features["labels"].int2str` function to get the class for a particular label index. Also note that, labels for test set are returned as -1 as they are missing. <details> <summary> Click here to see the full list of ImageNet class labels mapping: </summary> |id|Class| |--|-----| |0 | tench, Tinca tinca| |1 | goldfish, Carassius auratus| |2 | great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias| |3 | tiger shark, Galeocerdo cuvieri| |4 | hammerhead, hammerhead shark| |5 | electric ray, crampfish, numbfish, torpedo| |6 | stingray| |7 | cock| |8 | hen| |9 | ostrich, Struthio camelus| |10 | brambling, Fringilla montifringilla| |11 | goldfinch, Carduelis carduelis| |12 | house finch, linnet, Carpodacus mexicanus| |13 | junco, snowbird| |14 | indigo bunting, indigo finch, indigo bird, Passerina cyanea| |15 | robin, American robin, Turdus migratorius| |16 | bulbul| |17 | jay| |18 | magpie| |19 | chickadee| |20 | water ouzel, dipper| |21 | kite| |22 | bald eagle, American eagle, Haliaeetus leucocephalus| |23 | vulture| |24 | great grey owl, great gray owl, Strix nebulosa| |25 | European fire salamander, Salamandra salamandra| |26 | common newt, Triturus vulgaris| |27 | eft| |28 | spotted salamander, Ambystoma maculatum| |29 | axolotl, mud puppy, Ambystoma mexicanum| |30 | bullfrog, Rana catesbeiana| |31 | tree frog, tree-frog| |32 | tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui| |33 | loggerhead, loggerhead turtle, Caretta caretta| |34 | leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea| |35 | mud turtle| |36 | terrapin| |37 | box turtle, box tortoise| |38 | banded gecko| |39 | common iguana, iguana, Iguana iguana| |40 | American chameleon, anole, Anolis carolinensis| |41 | whiptail, whiptail lizard| |42 | agama| |43 | frilled lizard, Chlamydosaurus kingi| |44 | alligator lizard| |45 | Gila monster, Heloderma suspectum| |46 | green lizard, Lacerta viridis| |47 | African chameleon, Chamaeleo chamaeleon| |48 | Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis| |49 | African crocodile, Nile crocodile, Crocodylus niloticus| |50 | American alligator, Alligator mississipiensis| |51 | triceratops| |52 | thunder snake, worm snake, Carphophis amoenus| |53 | ringneck snake, ring-necked snake, ring snake| |54 | hognose snake, puff adder, sand viper| |55 | green snake, grass snake| |56 | king snake, kingsnake| |57 | garter snake, grass snake| |58 | water snake| |59 | vine snake| |60 | night snake, Hypsiglena torquata| |61 | boa constrictor, Constrictor constrictor| |62 | rock python, rock snake, Python sebae| |63 | Indian cobra, Naja naja| |64 | green mamba| |65 | sea snake| |66 | horned viper, cerastes, sand viper, horned asp, Cerastes cornutus| |67 | diamondback, diamondback rattlesnake, Crotalus adamanteus| |68 | sidewinder, horned rattlesnake, Crotalus cerastes| |69 | trilobite| |70 | harvestman, daddy longlegs, Phalangium opilio| |71 | scorpion| |72 | black and gold garden spider, Argiope aurantia| |73 | barn spider, Araneus cavaticus| |74 | garden spider, Aranea diademata| |75 | black widow, Latrodectus mactans| |76 | tarantula| |77 | wolf spider, hunting spider| |78 | tick| |79 | centipede| |80 | black grouse| |81 | ptarmigan| |82 | ruffed grouse, partridge, Bonasa umbellus| |83 | prairie chicken, prairie grouse, prairie fowl| |84 | peacock| |85 | quail| |86 | partridge| |87 | African grey, African gray, Psittacus erithacus| |88 | macaw| |89 | sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita| |90 | lorikeet| |91 | coucal| |92 | bee eater| |93 | hornbill| |94 | hummingbird| |95 | jacamar| |96 | toucan| |97 | drake| |98 | red-breasted merganser, Mergus serrator| |99 | goose| |100 | black swan, Cygnus atratus| |101 | tusker| |102 | echidna, spiny anteater, anteater| |103 | platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus| |104 | wallaby, brush kangaroo| |105 | koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus| |106 | wombat| |107 | jellyfish| |108 | sea anemone, anemone| |109 | brain coral| |110 | flatworm, platyhelminth| |111 | nematode, nematode worm, roundworm| |112 | conch| |113 | snail| |114 | slug| |115 | sea slug, nudibranch| |116 | chiton, coat-of-mail shell, sea cradle, polyplacophore| |117 | chambered nautilus, pearly nautilus, nautilus| |118 | Dungeness crab, Cancer magister| |119 | rock crab, Cancer irroratus| |120 | fiddler crab| |121 | king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica| |122 | American lobster, Northern lobster, Maine lobster, Homarus americanus| |123 | spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish| |124 | crayfish, crawfish, crawdad, crawdaddy| |125 | hermit crab| |126 | isopod| |127 | white stork, Ciconia ciconia| |128 | black stork, Ciconia nigra| |129 | spoonbill| |130 | flamingo| |131 | little blue heron, Egretta caerulea| |132 | American egret, great white heron, Egretta albus| |133 | bittern| |134 | crane| |135 | limpkin, Aramus pictus| |136 | European gallinule, Porphyrio porphyrio| |137 | American coot, marsh hen, mud hen, water hen, Fulica americana| |138 | bustard| |139 | ruddy turnstone, Arenaria interpres| |140 | red-backed sandpiper, dunlin, Erolia alpina| |141 | redshank, Tringa totanus| |142 | dowitcher| |143 | oystercatcher, oyster catcher| |144 | pelican| |145 | king penguin, Aptenodytes patagonica| |146 | albatross, mollymawk| |147 | grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus| |148 | killer whale, killer, orca, grampus, sea wolf, Orcinus orca| |149 | dugong, Dugong dugon| |150 | sea lion| |151 | Chihuahua| |152 | Japanese spaniel| |153 | Maltese dog, Maltese terrier, Maltese| |154 | Pekinese, Pekingese, Peke| |155 | Shih-Tzu| |156 | Blenheim spaniel| |157 | papillon| |158 | toy terrier| |159 | Rhodesian ridgeback| |160 | Afghan hound, Afghan| |161 | basset, basset hound| |162 | beagle| |163 | bloodhound, sleuthhound| |164 | bluetick| |165 | black-and-tan coonhound| |166 | Walker hound, Walker foxhound| |167 | English foxhound| |168 | redbone| |169 | borzoi, Russian wolfhound| |170 | Irish wolfhound| |171 | Italian greyhound| |172 | whippet| |173 | Ibizan hound, Ibizan Podenco| |174 | Norwegian elkhound, elkhound| |175 | otterhound, otter hound| |176 | Saluki, gazelle hound| |177 | Scottish deerhound, deerhound| |178 | Weimaraner| |179 | Staffordshire bullterrier, Staffordshire bull terrier| |180 | American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier| |181 | Bedlington terrier| |182 | Border terrier| |183 | Kerry blue terrier| |184 | Irish terrier| |185 | Norfolk terrier| |186 | Norwich terrier| |187 | Yorkshire terrier| |188 | wire-haired fox terrier| |189 | Lakeland terrier| |190 | Sealyham terrier, Sealyham| |191 | Airedale, Airedale terrier| |192 | cairn, cairn terrier| |193 | Australian terrier| |194 | Dandie Dinmont, Dandie Dinmont terrier| |195 | Boston bull, Boston terrier| |196 | miniature schnauzer| |197 | giant schnauzer| |198 | standard schnauzer| |199 | Scotch terrier, Scottish terrier, Scottie| |200 | Tibetan terrier, chrysanthemum dog| |201 | silky terrier, Sydney silky| |202 | soft-coated wheaten terrier| |203 | West Highland white terrier| |204 | Lhasa, Lhasa apso| |205 | flat-coated retriever| |206 | curly-coated retriever| |207 | golden retriever| |208 | Labrador retriever| |209 | Chesapeake Bay retriever| |210 | German short-haired pointer| |211 | vizsla, Hungarian pointer| |212 | English setter| |213 | Irish setter, red setter| |214 | Gordon setter| |215 | Brittany spaniel| |216 | clumber, clumber spaniel| |217 | English springer, English springer spaniel| |218 | Welsh springer spaniel| |219 | cocker spaniel, English cocker spaniel, cocker| |220 | Sussex spaniel| |221 | Irish water spaniel| |222 | kuvasz| |223 | schipperke| |224 | groenendael| |225 | malinois| |226 | briard| |227 | kelpie| |228 | komondor| |229 | Old English sheepdog, bobtail| |230 | Shetland sheepdog, Shetland sheep dog, Shetland| |231 | collie| |232 | Border collie| |233 | Bouvier des Flandres, Bouviers des Flandres| |234 | Rottweiler| |235 | German shepherd, German shepherd dog, German police dog, alsatian| |236 | Doberman, Doberman pinscher| |237 | miniature pinscher| |238 | Greater Swiss Mountain dog| |239 | Bernese mountain dog| |240 | Appenzeller| |241 | EntleBucher| |242 | boxer| |243 | bull mastiff| |244 | Tibetan mastiff| |245 | French bulldog| |246 | Great Dane| |247 | Saint Bernard, St Bernard| |248 | Eskimo dog, husky| |249 | malamute, malemute, Alaskan malamute| |250 | Siberian husky| |251 | dalmatian, coach dog, carriage dog| |252 | affenpinscher, monkey pinscher, monkey dog| |253 | basenji| |254 | pug, pug-dog| |255 | Leonberg| |256 | Newfoundland, Newfoundland dog| |257 | Great Pyrenees| |258 | Samoyed, Samoyede| |259 | Pomeranian| |260 | chow, chow chow| |261 | keeshond| |262 | Brabancon griffon| |263 | Pembroke, Pembroke Welsh corgi| |264 | Cardigan, Cardigan Welsh corgi| |265 | toy poodle| |266 | miniature poodle| |267 | standard poodle| |268 | Mexican hairless| |269 | timber wolf, grey wolf, gray wolf, Canis lupus| |270 | white wolf, Arctic wolf, Canis lupus tundrarum| |271 | red wolf, maned wolf, Canis rufus, Canis niger| |272 | coyote, prairie wolf, brush wolf, Canis latrans| |273 | dingo, warrigal, warragal, Canis dingo| |274 | dhole, Cuon alpinus| |275 | African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus| |276 | hyena, hyaena| |277 | red fox, Vulpes vulpes| |278 | kit fox, Vulpes macrotis| |279 | Arctic fox, white fox, Alopex lagopus| |280 | grey fox, gray fox, Urocyon cinereoargenteus| |281 | tabby, tabby cat| |282 | tiger cat| |283 | Persian cat| |284 | Siamese cat, Siamese| |285 | Egyptian cat| |286 | cougar, puma, catamount, mountain lion, painter, panther, Felis concolor| |287 | lynx, catamount| |288 | leopard, Panthera pardus| |289 | snow leopard, ounce, Panthera uncia| |290 | jaguar, panther, Panthera onca, Felis onca| |291 | lion, king of beasts, Panthera leo| |292 | tiger, Panthera tigris| |293 | cheetah, chetah, Acinonyx jubatus| |294 | brown bear, bruin, Ursus arctos| |295 | American black bear, black bear, Ursus americanus, Euarctos americanus| |296 | ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus| |297 | sloth bear, Melursus ursinus, Ursus ursinus| |298 | mongoose| |299 | meerkat, mierkat| |300 | tiger beetle| |301 | ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle| |302 | ground beetle, carabid beetle| |303 | long-horned beetle, longicorn, longicorn beetle| |304 | leaf beetle, chrysomelid| |305 | dung beetle| |306 | rhinoceros beetle| |307 | weevil| |308 | fly| |309 | bee| |310 | ant, emmet, pismire| |311 | grasshopper, hopper| |312 | cricket| |313 | walking stick, walkingstick, stick insect| |314 | cockroach, roach| |315 | mantis, mantid| |316 | cicada, cicala| |317 | leafhopper| |318 | lacewing, lacewing fly| |319 | dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk| |320 | damselfly| |321 | admiral| |322 | ringlet, ringlet butterfly| |323 | monarch, monarch butterfly, milkweed butterfly, Danaus plexippus| |324 | cabbage butterfly| |325 | sulphur butterfly, sulfur butterfly| |326 | lycaenid, lycaenid butterfly| |327 | starfish, sea star| |328 | sea urchin| |329 | sea cucumber, holothurian| |330 | wood rabbit, cottontail, cottontail rabbit| |331 | hare| |332 | Angora, Angora rabbit| |333 | hamster| |334 | porcupine, hedgehog| |335 | fox squirrel, eastern fox squirrel, Sciurus niger| |336 | marmot| |337 | beaver| |338 | guinea pig, Cavia cobaya| |339 | sorrel| |340 | zebra| |341 | hog, pig, grunter, squealer, Sus scrofa| |342 | wild boar, boar, Sus scrofa| |343 | warthog| |344 | hippopotamus, hippo, river horse, Hippopotamus amphibius| |345 | ox| |346 | water buffalo, water ox, Asiatic buffalo, Bubalus bubalis| |347 | bison| |348 | ram, tup| |349 | bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis| |350 | ibex, Capra ibex| |351 | hartebeest| |352 | impala, Aepyceros melampus| |353 | gazelle| |354 | Arabian camel, dromedary, Camelus dromedarius| |355 | llama| |356 | weasel| |357 | mink| |358 | polecat, fitch, foulmart, foumart, Mustela putorius| |359 | black-footed ferret, ferret, Mustela nigripes| |360 | otter| |361 | skunk, polecat, wood pussy| |362 | badger| |363 | armadillo| |364 | three-toed sloth, ai, Bradypus tridactylus| |365 | orangutan, orang, orangutang, Pongo pygmaeus| |366 | gorilla, Gorilla gorilla| |367 | chimpanzee, chimp, Pan troglodytes| |368 | gibbon, Hylobates lar| |369 | siamang, Hylobates syndactylus, Symphalangus syndactylus| |370 | guenon, guenon monkey| |371 | patas, hussar monkey, Erythrocebus patas| |372 | baboon| |373 | macaque| |374 | langur| |375 | colobus, colobus monkey| |376 | proboscis monkey, Nasalis larvatus| |377 | marmoset| |378 | capuchin, ringtail, Cebus capucinus| |379 | howler monkey, howler| |380 | titi, titi monkey| |381 | spider monkey, Ateles geoffroyi| |382 | squirrel monkey, Saimiri sciureus| |383 | Madagascar cat, ring-tailed lemur, Lemur catta| |384 | indri, indris, Indri indri, Indri brevicaudatus| |385 | Indian elephant, Elephas maximus| |386 | African elephant, Loxodonta africana| |387 | lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens| |388 | giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca| |389 | barracouta, snoek| |390 | eel| |391 | coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch| |392 | rock beauty, Holocanthus tricolor| |393 | anemone fish| |394 | sturgeon| |395 | gar, garfish, garpike, billfish, Lepisosteus osseus| |396 | lionfish| |397 | puffer, pufferfish, blowfish, globefish| |398 | abacus| |399 | abaya| |400 | academic gown, academic robe, judge's robe| |401 | accordion, piano accordion, squeeze box| |402 | acoustic guitar| |403 | aircraft carrier, carrier, flattop, attack aircraft carrier| |404 | airliner| |405 | airship, dirigible| |406 | altar| |407 | ambulance| |408 | amphibian, amphibious vehicle| |409 | analog clock| |410 | apiary, bee house| |411 | apron| |412 | ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin| |413 | assault rifle, assault gun| |414 | backpack, back pack, knapsack, packsack, rucksack, haversack| |415 | bakery, bakeshop, bakehouse| |416 | balance beam, beam| |417 | balloon| |418 | ballpoint, ballpoint pen, ballpen, Biro| |419 | Band Aid| |420 | banjo| |421 | bannister, banister, balustrade, balusters, handrail| |422 | barbell| |423 | barber chair| |424 | barbershop| |425 | barn| |426 | barometer| |427 | barrel, cask| |428 | barrow, garden cart, lawn cart, wheelbarrow| |429 | baseball| |430 | basketball| |431 | bassinet| |432 | bassoon| |433 | bathing cap, swimming cap| |434 | bath towel| |435 | bathtub, bathing tub, bath, tub| |436 | beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon| |437 | beacon, lighthouse, beacon light, pharos| |438 | beaker| |439 | bearskin, busby, shako| |440 | beer bottle| |441 | beer glass| |442 | bell cote, bell cot| |443 | bib| |444 | bicycle-built-for-two, tandem bicycle, tandem| |445 | bikini, two-piece| |446 | binder, ring-binder| |447 | binoculars, field glasses, opera glasses| |448 | birdhouse| |449 | boathouse| |450 | bobsled, bobsleigh, bob| |451 | bolo tie, bolo, bola tie, bola| |452 | bonnet, poke bonnet| |453 | bookcase| |454 | bookshop, bookstore, bookstall| |455 | bottlecap| |456 | bow| |457 | bow tie, bow-tie, bowtie| |458 | brass, memorial tablet, plaque| |459 | brassiere, bra, bandeau| |460 | breakwater, groin, groyne, mole, bulwark, seawall, jetty| |461 | breastplate, aegis, egis| |462 | broom| |463 | bucket, pail| |464 | buckle| |465 | bulletproof vest| |466 | bullet train, bullet| |467 | butcher shop, meat market| |468 | cab, hack, taxi, taxicab| |469 | caldron, cauldron| |470 | candle, taper, wax light| |471 | cannon| |472 | canoe| |473 | can opener, tin opener| |474 | cardigan| |475 | car mirror| |476 | carousel, carrousel, merry-go-round, roundabout, whirligig| |477 | carpenter's kit, tool kit| |478 | carton| |479 | car wheel| |480 | cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM| |481 | cassette| |482 | cassette player| |483 | castle| |484 | catamaran| |485 | CD player| |486 | cello, violoncello| |487 | cellular telephone, cellular phone, cellphone, cell, mobile phone| |488 | chain| |489 | chainlink fence| |490 | chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour| |491 | chain saw, chainsaw| |492 | chest| |493 | chiffonier, commode| |494 | chime, bell, gong| |495 | china cabinet, china closet| |496 | Christmas stocking| |497 | church, church building| |498 | cinema, movie theater, movie theatre, movie house, picture palace| |499 | cleaver, meat cleaver, chopper| |500 | cliff dwelling| |501 | cloak| |502 | clog, geta, patten, sabot| |503 | cocktail shaker| |504 | coffee mug| |505 | coffeepot| |506 | coil, spiral, volute, whorl, helix| |507 | combination lock| |508 | computer keyboard, keypad| |509 | confectionery, confectionary, candy store| |510 | container ship, containership, container vessel| |511 | convertible| |512 | corkscrew, bottle screw| |513 | cornet, horn, trumpet, trump| |514 | cowboy boot| |515 | cowboy hat, ten-gallon hat| |516 | cradle| |517 | crane_1| |518 | crash helmet| |519 | crate| |520 | crib, cot| |521 | Crock Pot| |522 | croquet ball| |523 | crutch| |524 | cuirass| |525 | dam, dike, dyke| |526 | desk| |527 | desktop computer| |528 | dial telephone, dial phone| |529 | diaper, nappy, napkin| |530 | digital clock| |531 | digital watch| |532 | dining table, board| |533 | dishrag, dishcloth| |534 | dishwasher, dish washer, dishwashing machine| |535 | disk brake, disc brake| |536 | dock, dockage, docking facility| |537 | dogsled, dog sled, dog sleigh| |538 | dome| |539 | doormat, welcome mat| |540 | drilling platform, offshore rig| |541 | drum, membranophone, tympan| |542 | drumstick| |543 | dumbbell| |544 | Dutch oven| |545 | electric fan, blower| |546 | electric guitar| |547 | electric locomotive| |548 | entertainment center| |549 | envelope| |550 | espresso maker| |551 | face powder| |552 | feather boa, boa| |553 | file, file cabinet, filing cabinet| |554 | fireboat| |555 | fire engine, fire truck| |556 | fire screen, fireguard| |557 | flagpole, flagstaff| |558 | flute, transverse flute| |559 | folding chair| |560 | football helmet| |561 | forklift| |562 | fountain| |563 | fountain pen| |564 | four-poster| |565 | freight car| |566 | French horn, horn| |567 | frying pan, frypan, skillet| |568 | fur coat| |569 | garbage truck, dustcart| |570 | gasmask, respirator, gas helmet| |571 | gas pump, gasoline pump, petrol pump, island dispenser| |572 | goblet| |573 | go-kart| |574 | golf ball| |575 | golfcart, golf cart| |576 | gondola| |577 | gong, tam-tam| |578 | gown| |579 | grand piano, grand| |580 | greenhouse, nursery, glasshouse| |581 | grille, radiator grille| |582 | grocery store, grocery, food market, market| |583 | guillotine| |584 | hair slide| |585 | hair spray| |586 | half track| |587 | hammer| |588 | hamper| |589 | hand blower, blow dryer, blow drier, hair dryer, hair drier| |590 | hand-held computer, hand-held microcomputer| |591 | handkerchief, hankie, hanky, hankey| |592 | hard disc, hard disk, fixed disk| |593 | harmonica, mouth organ, harp, mouth harp| |594 | harp| |595 | harvester, reaper| |596 | hatchet| |597 | holster| |598 | home theater, home theatre| |599 | honeycomb| |600 | hook, claw| |601 | hoopskirt, crinoline| |602 | horizontal bar, high bar| |603 | horse cart, horse-cart| |604 | hourglass| |605 | iPod| |606 | iron, smoothing iron| |607 | jack-o'-lantern| |608 | jean, blue jean, denim| |609 | jeep, landrover| |610 | jersey, T-shirt, tee shirt| |611 | jigsaw puzzle| |612 | jinrikisha, ricksha, rickshaw| |613 | joystick| |614 | kimono| |615 | knee pad| |616 | knot| |617 | lab coat, laboratory coat| |618 | ladle| |619 | lampshade, lamp shade| |620 | laptop, laptop computer| |621 | lawn mower, mower| |622 | lens cap, lens cover| |623 | letter opener, paper knife, paperknife| |624 | library| |625 | lifeboat| |626 | lighter, light, igniter, ignitor| |627 | limousine, limo| |628 | liner, ocean liner| |629 | lipstick, lip rouge| |630 | Loafer| |631 | lotion| |632 | loudspeaker, speaker, speaker unit, loudspeaker system, speaker system| |633 | loupe, jeweler's loupe| |634 | lumbermill, sawmill| |635 | magnetic compass| |636 | mailbag, postbag| |637 | mailbox, letter box| |638 | maillot| |639 | maillot, tank suit| |640 | manhole cover| |641 | maraca| |642 | marimba, xylophone| |643 | mask| |644 | matchstick| |645 | maypole| |646 | maze, labyrinth| |647 | measuring cup| |648 | medicine chest, medicine cabinet| |649 | megalith, megalithic structure| |650 | microphone, mike| |651 | microwave, microwave oven| |652 | military uniform| |653 | milk can| |654 | minibus| |655 | miniskirt, mini| |656 | minivan| |657 | missile| |658 | mitten| |659 | mixing bowl| |660 | mobile home, manufactured home| |661 | Model T| |662 | modem| |663 | monastery| |664 | monitor| |665 | moped| |666 | mortar| |667 | mortarboard| |668 | mosque| |669 | mosquito net| |670 | motor scooter, scooter| |671 | mountain bike, all-terrain bike, off-roader| |672 | mountain tent| |673 | mouse, computer mouse| |674 | mousetrap| |675 | moving van| |676 | muzzle| |677 | nail| |678 | neck brace| |679 | necklace| |680 | nipple| |681 | notebook, notebook computer| |682 | obelisk| |683 | oboe, hautboy, hautbois| |684 | ocarina, sweet potato| |685 | odometer, hodometer, mileometer, milometer| |686 | oil filter| |687 | organ, pipe organ| |688 | oscilloscope, scope, cathode-ray oscilloscope, CRO| |689 | overskirt| |690 | oxcart| |691 | oxygen mask| |692 | packet| |693 | paddle, boat paddle| |694 | paddlewheel, paddle wheel| |695 | padlock| |696 | paintbrush| |697 | pajama, pyjama, pj's, jammies| |698 | palace| |699 | panpipe, pandean pipe, syrinx| |700 | paper towel| |701 | parachute, chute| |702 | parallel bars, bars| |703 | park bench| |704 | parking meter| |705 | passenger car, coach, carriage| |706 | patio, terrace| |707 | pay-phone, pay-station| |708 | pedestal, plinth, footstall| |709 | pencil box, pencil case| |710 | pencil sharpener| |711 | perfume, essence| |712 | Petri dish| |713 | photocopier| |714 | pick, plectrum, plectron| |715 | pickelhaube| |716 | picket fence, paling| |717 | pickup, pickup truck| |718 | pier| |719 | piggy bank, penny bank| |720 | pill bottle| |721 | pillow| |722 | ping-pong ball| |723 | pinwheel| |724 | pirate, pirate ship| |725 | pitcher, ewer| |726 | plane, carpenter's plane, woodworking plane| |727 | planetarium| |728 | plastic bag| |729 | plate rack| |730 | plow, plough| |731 | plunger, plumber's helper| |732 | Polaroid camera, Polaroid Land camera| |733 | pole| |734 | police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria| |735 | poncho| |736 | pool table, billiard table, snooker table| |737 | pop bottle, soda bottle| |738 | pot, flowerpot| |739 | potter's wheel| |740 | power drill| |741 | prayer rug, prayer mat| |742 | printer| |743 | prison, prison house| |744 | projectile, missile| |745 | projector| |746 | puck, hockey puck| |747 | punching bag, punch bag, punching ball, punchball| |748 | purse| |749 | quill, quill pen| |750 | quilt, comforter, comfort, puff| |751 | racer, race car, racing car| |752 | racket, racquet| |753 | radiator| |754 | radio, wireless| |755 | radio telescope, radio reflector| |756 | rain barrel| |757 | recreational vehicle, RV, R.V.| |758 | reel| |759 | reflex camera| |760 | refrigerator, icebox| |761 | remote control, remote| |762 | restaurant, eating house, eating place, eatery| |763 | revolver, six-gun, six-shooter| |764 | rifle| |765 | rocking chair, rocker| |766 | rotisserie| |767 | rubber eraser, rubber, pencil eraser| |768 | rugby ball| |769 | rule, ruler| |770 | running shoe| |771 | safe| |772 | safety pin| |773 | saltshaker, salt shaker| |774 | sandal| |775 | sarong| |776 | sax, saxophone| |777 | scabbard| |778 | scale, weighing machine| |779 | school bus| |780 | schooner| |781 | scoreboard| |782 | screen, CRT screen| |783 | screw| |784 | screwdriver| |785 | seat belt, seatbelt| |786 | sewing machine| |787 | shield, buckler| |788 | shoe shop, shoe-shop, shoe store| |789 | shoji| |790 | shopping basket| |791 | shopping cart| |792 | shovel| |793 | shower cap| |794 | shower curtain| |795 | ski| |796 | ski mask| |797 | sleeping bag| |798 | slide rule, slipstick| |799 | sliding door| |800 | slot, one-armed bandit| |801 | snorkel| |802 | snowmobile| |803 | snowplow, snowplough| |804 | soap dispenser| |805 | soccer ball| |806 | sock| |807 | solar dish, solar collector, solar furnace| |808 | sombrero| |809 | soup bowl| |810 | space bar| |811 | space heater| |812 | space shuttle| |813 | spatula| |814 | speedboat| |815 | spider web, spider's web| |816 | spindle| |817 | sports car, sport car| |818 | spotlight, spot| |819 | stage| |820 | steam locomotive| |821 | steel arch bridge| |822 | steel drum| |823 | stethoscope| |824 | stole| |825 | stone wall| |826 | stopwatch, stop watch| |827 | stove| |828 | strainer| |829 | streetcar, tram, tramcar, trolley, trolley car| |830 | stretcher| |831 | studio couch, day bed| |832 | stupa, tope| |833 | submarine, pigboat, sub, U-boat| |834 | suit, suit of clothes| |835 | sundial| |836 | sunglass| |837 | sunglasses, dark glasses, shades| |838 | sunscreen, sunblock, sun blocker| |839 | suspension bridge| |840 | swab, swob, mop| |841 | sweatshirt| |842 | swimming trunks, bathing trunks| |843 | swing| |844 | switch, electric switch, electrical switch| |845 | syringe| |846 | table lamp| |847 | tank, army tank, armored combat vehicle, armoured combat vehicle| |848 | tape player| |849 | teapot| |850 | teddy, teddy bear| |851 | television, television system| |852 | tennis ball| |853 | thatch, thatched roof| |854 | theater curtain, theatre curtain| |855 | thimble| |856 | thresher, thrasher, threshing machine| |857 | throne| |858 | tile roof| |859 | toaster| |860 | tobacco shop, tobacconist shop, tobacconist| |861 | toilet seat| |862 | torch| |863 | totem pole| |864 | tow truck, tow car, wrecker| |865 | toyshop| |866 | tractor| |867 | trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi| |868 | tray| |869 | trench coat| |870 | tricycle, trike, velocipede| |871 | trimaran| |872 | tripod| |873 | triumphal arch| |874 | trolleybus, trolley coach, trackless trolley| |875 | trombone| |876 | tub, vat| |877 | turnstile| |878 | typewriter keyboard| |879 | umbrella| |880 | unicycle, monocycle| |881 | upright, upright piano| |882 | vacuum, vacuum cleaner| |883 | vase| |884 | vault| |885 | velvet| |886 | vending machine| |887 | vestment| |888 | viaduct| |889 | violin, fiddle| |890 | volleyball| |891 | waffle iron| |892 | wall clock| |893 | wallet, billfold, notecase, pocketbook| |894 | wardrobe, closet, press| |895 | warplane, military plane| |896 | washbasin, handbasin, washbowl, lavabo, wash-hand basin| |897 | washer, automatic washer, washing machine| |898 | water bottle| |899 | water jug| |900 | water tower| |901 | whiskey jug| |902 | whistle| |903 | wig| |904 | window screen| |905 | window shade| |906 | Windsor tie| |907 | wine bottle| |908 | wing| |909 | wok| |910 | wooden spoon| |911 | wool, woolen, woollen| |912 | worm fence, snake fence, snake-rail fence, Virginia fence| |913 | wreck| |914 | yawl| |915 | yurt| |916 | web site, website, internet site, site| |917 | comic book| |918 | crossword puzzle, crossword| |919 | street sign| |920 | traffic light, traffic signal, stoplight| |921 | book jacket, dust cover, dust jacket, dust wrapper| |922 | menu| |923 | plate| |924 | guacamole| |925 | consomme| |926 | hot pot, hotpot| |927 | trifle| |928 | ice cream, icecream| |929 | ice lolly, lolly, lollipop, popsicle| |930 | French loaf| |931 | bagel, beigel| |932 | pretzel| |933 | cheeseburger| |934 | hotdog, hot dog, red hot| |935 | mashed potato| |936 | head cabbage| |937 | broccoli| |938 | cauliflower| |939 | zucchini, courgette| |940 | spaghetti squash| |941 | acorn squash| |942 | butternut squash| |943 | cucumber, cuke| |944 | artichoke, globe artichoke| |945 | bell pepper| |946 | cardoon| |947 | mushroom| |948 | Granny Smith| |949 | strawberry| |950 | orange| |951 | lemon| |952 | fig| |953 | pineapple, ananas| |954 | banana| |955 | jackfruit, jak, jack| |956 | custard apple| |957 | pomegranate| |958 | hay| |959 | carbonara| |960 | chocolate sauce, chocolate syrup| |961 | dough| |962 | meat loaf, meatloaf| |963 | pizza, pizza pie| |964 | potpie| |965 | burrito| |966 | red wine| |967 | espresso| |968 | cup| |969 | eggnog| |970 | alp| |971 | bubble| |972 | cliff, drop, drop-off| |973 | coral reef| |974 | geyser| |975 | lakeside, lakeshore| |976 | promontory, headland, head, foreland| |977 | sandbar, sand bar| |978 | seashore, coast, seacoast, sea-coast| |979 | valley, vale| |980 | volcano| |981 | ballplayer, baseball player| |982 | groom, bridegroom| |983 | scuba diver| |984 | rapeseed| |985 | daisy| |986 | yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum| |987 | corn| |988 | acorn| |989 | hip, rose hip, rosehip| |990 | buckeye, horse chestnut, conker| |991 | coral fungus| |992 | agaric| |993 | gyromitra| |994 | stinkhorn, carrion fungus| |995 | earthstar| |996 | hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa| |997 | bolete| |998 | ear, spike, capitulum| |999 | toilet tissue, toilet paper, bathroom tissue| </details> ### Data Splits | |train |validation| test | |-------------|------:|---------:|------:| |# of examples|1281167|50000 |100000 | ## Dataset Creation ### Curation Rationale The ImageNet project was inspired by two important needs in computer vision research. The first was the need to establish a clear North Star problem in computer vision. While the field enjoyed an abundance of important tasks to work on, from stereo vision to image retrieval, from 3D reconstruction to image segmentation, object categorization was recognized to be one of the most fundamental capabilities of both human and machine vision. Hence there was a growing demand for a high quality object categorization benchmark with clearly established evaluation metrics. Second, there was a critical need for more data to enable more generalizable machine learning methods. Ever since the birth of the digital era and the availability of web-scale data exchanges, researchers in these fields have been working hard to design more and more sophisticated algorithms to index, retrieve, organize and annotate multimedia data. But good research requires good resources. To tackle this problem at scale (think of your growing personal collection of digital images, or videos, or a commercial web search engine’s database), it was critical to provide researchers with a large-scale image database for both training and testing. The convergence of these two intellectual reasons motivated us to build ImageNet. ### Source Data #### Initial Data Collection and Normalization Initial data for ImageNet image classification task consists of photographs collected from [Flickr](https://www.flickr.com) and other search engines, manually labeled with the presence of one of 1000 object categories. Constructing ImageNet was an effort to scale up an image classification dataset to cover most nouns in English using tens of millions of manually verified photographs [1](https://ieeexplore.ieee.org/abstract/document/5206848). The image classification task of ILSVRC came as a direct extension of this effort. A subset of categories and images was chosen and fixed to provide a standardized benchmark while the rest of ImageNet continued to grow. #### Who are the source language producers? WordNet synsets further quality controlled by human annotators. The images are from Flickr. ### Annotations #### Annotation process The annotation process of collecting ImageNet for image classification task is a three step process. 1. Defining the 1000 object categories for the image classification task. These categories have evolved over the years. 1. Collecting the candidate image for these object categories using a search engine. 1. Quality control on the candidate images by using human annotators on Amazon Mechanical Turk (AMT) to make sure the image has the synset it was collected for. See the section 3.1 in [1](https://arxiv.org/abs/1409.0575) for more details on data collection procedure and [2](https://ieeexplore.ieee.org/abstract/document/5206848) for general information on ImageNet. #### Who are the annotators? Images are automatically fetched from an image search engine based on the synsets and filtered using human annotators on Amazon Mechanical Turk. See [1](https://arxiv.org/abs/1409.0575) for more details. ### Personal and Sensitive Information The 1,000 categories selected for this subset contain only 3 people categories (scuba diver, bridegroom, and baseball player) while the full ImageNet contains 2,832 people categories under the person subtree (accounting for roughly 8.3% of the total images). This subset does contain the images of people without their consent. Though, the study in [[1]](https://image-net.org/face-obfuscation/) on obfuscating faces of the people in the ImageNet 2012 subset shows that blurring people's faces causes a very minor decrease in accuracy (~0.6%) suggesting that privacy-aware models can be trained on ImageNet. On larger ImageNet, there has been [an attempt](https://arxiv.org/abs/1912.07726) at filtering and balancing the people subtree in the larger ImageNet. ## Considerations for Using the Data ### Social Impact of Dataset The ImageNet dataset has been very crucial in advancement of deep learning technology as being the standard benchmark for the computer vision models. The dataset aims to probe models on their understanding of the objects and has become the de-facto dataset for this purpose. ImageNet is still one of the major datasets on which models are evaluated for their generalization in computer vision capabilities as the field moves towards self-supervised algorithms. Please see the future section in [1](https://arxiv.org/abs/1409.0575) for a discussion on social impact of the dataset. ### Discussion of Biases 1. A [study](https://image-net.org/update-sep-17-2019.php) of the history of the multiple layers (taxonomy, object classes and labeling) of ImageNet and WordNet in 2019 described how bias is deeply embedded in most classification approaches for of all sorts of images. 1. A [study](https://arxiv.org/abs/1811.12231) has also shown that ImageNet trained models are biased towards texture rather than shapes which in contrast with how humans do object classification. Increasing the shape bias improves the accuracy and robustness. 1. Another [study](https://arxiv.org/abs/2109.13228) more potential issues and biases with the ImageNet dataset and provides an alternative benchmark for image classification task. The data collected contains humans without their consent. 1. ImageNet data with face obfuscation is also provided at [this link](https://image-net.org/face-obfuscation/) 1. A study on genealogy of ImageNet is can be found at [this link](https://journals.sagepub.com/doi/full/10.1177/20539517211035955) about the "norms, values, and assumptions" in ImageNet. 1. See [this study](https://arxiv.org/abs/1912.07726) on filtering and balancing the distribution of people subtree in the larger complete ImageNet. ### Other Known Limitations 1. Since most of the images were collected from internet, keep in mind that some images in ImageNet might be subject to copyrights. See the following papers for more details: [[1]](https://arxiv.org/abs/2109.13228) [[2]](https://arxiv.org/abs/1409.0575) [[3]](https://ieeexplore.ieee.org/abstract/document/5206848). ## Additional Information ### Dataset Curators Authors of [[1]](https://arxiv.org/abs/1409.0575) and [[2]](https://ieeexplore.ieee.org/abstract/document/5206848): - Olga Russakovsky - Jia Deng - Hao Su - Jonathan Krause - Sanjeev Satheesh - Wei Dong - Richard Socher - Li-Jia Li - Kai Li - Sean Ma - Zhiheng Huang - Andrej Karpathy - Aditya Khosla - Michael Bernstein - Alexander C Berg - Li Fei-Fei ### Licensing Information In exchange for permission to use the ImageNet database (the "Database") at Princeton University and Stanford University, Researcher hereby agrees to the following terms and conditions: 1. Researcher shall use the Database only for non-commercial research and educational purposes. 1. Princeton University and Stanford University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. 1. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the ImageNet team, Princeton University, and Stanford University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted images that he or she may create from the Database. 1. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. 1. Princeton University and Stanford University reserve the right to terminate Researcher's access to the Database at any time. 1. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer. 1. The law of the State of New Jersey shall apply to all disputes under this agreement. ### Citation Information ```bibtex @article{imagenet15russakovsky, Author = {Olga Russakovsky and Jia Deng and Hao Su and Jonathan Krause and Sanjeev Satheesh and Sean Ma and Zhiheng Huang and Andrej Karpathy and Aditya Khosla and Michael Bernstein and Alexander C. Berg and Li Fei-Fei}, Title = { {ImageNet Large Scale Visual Recognition Challenge} }, Year = {2015}, journal = {International Journal of Computer Vision (IJCV)}, doi = {10.1007/s11263-015-0816-y}, volume={115}, number={3}, pages={211-252} } ``` ### Contributions Thanks to [@apsdehal](https://github.com/apsdehal) for adding this dataset.
[ "BEAR" ]
SilasK/mistral-7b-medqa-version3
SilasK
null
[ "peft", "tensorboard", "safetensors", "mistral", "trl", "sft", "generated_from_trainer", "base_model:mistralai/Mistral-7B-Instruct-v0.1", "base_model:adapter:mistralai/Mistral-7B-Instruct-v0.1", "license:apache-2.0", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-09T15:23:58Z"
2024-03-09T15:29:54+00:00
0
0
--- base_model: mistralai/Mistral-7B-Instruct-v0.1 library_name: peft license: apache-2.0 tags: - trl - sft - generated_from_trainer model-index: - name: mistral-7b-medqa-version3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-medqa-version3 This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 1e-08 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
SilasK/llama-7b-medqa_version_3
SilasK
null
[ "peft", "tensorboard", "safetensors", "llama", "trl", "sft", "generated_from_trainer", "base_model:huggyllama/llama-7b", "base_model:adapter:huggyllama/llama-7b", "license:other", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-09T15:53:00Z"
2024-03-09T15:57:58+00:00
0
0
--- base_model: huggyllama/llama-7b library_name: peft license: other tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-medqa_version_3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-medqa_version_3 This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 1e-11 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
SilasK/llama-7b-medqa_version_5
SilasK
null
[ "peft", "tensorboard", "safetensors", "llama", "trl", "sft", "generated_from_trainer", "base_model:huggyllama/llama-7b", "base_model:adapter:huggyllama/llama-7b", "license:other", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-11T18:30:44Z"
2024-03-12T05:21:27+00:00
0
0
--- base_model: huggyllama/llama-7b library_name: peft license: other tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-medqa_version_5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-medqa_version_5 This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 2 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
SilasK/llama-7b-medqa_version_6
SilasK
null
[ "peft", "tensorboard", "safetensors", "llama", "trl", "sft", "generated_from_trainer", "base_model:huggyllama/llama-7b", "base_model:adapter:huggyllama/llama-7b", "license:other", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-12T17:17:57Z"
2024-03-13T08:31:59+00:00
0
0
--- base_model: huggyllama/llama-7b library_name: peft license: other tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-medqa_version_6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-medqa_version_6 This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 4 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
ostapeno/library-mistral7B_flan_5ep_sharedA
ostapeno
null
[ "region:us" ]
"2024-03-12T23:16:50Z"
2024-03-15T17:41:32+00:00
0
0
--- {} --- Number of experts present in the library: 256 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | aeslc_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | race_middle_Is_this_the_right_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | true_case | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/true_case | lora | | cot_strategyqa_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | quarel_do_not_use | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_do_not_use | lora | | word_segment | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | stream_aqua_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_aqua_ii | lora | | wiki_hop_original_choose_best_object_affirmative_3 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | duorc_ParaphraseRC_movie_director | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | math_dataset_algebra__linear_1d_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cot_gsm8k_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | social_i_qa_Show_choices_and_generate_index | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dream_generate_last_utterance | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | quail_context_question_description_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | cot_creak_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_creak_ii | lora | | ropes_background_new_situation_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | cot_esnli | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_esnli | lora | | anli_r3_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | adversarial_qa_droberta_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | cos_e_v1_11_i_think | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | gem_wiki_lingua_english_en_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | glue_qqp_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | wiki_qa_Topic_Prediction_Answer_Only | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | cot_creak | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_creak | lora | | trivia_qa_rc_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | duorc_SelfRC_title_generation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | wiqa_what_is_the_final_step_of_the_following_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | glue_mnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | quail_context_question_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | adversarial_qa_dbidaf_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | adversarial_qa_droberta_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | cos_e_v1_11_question_description_option_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | wiki_hop_original_generate_subject_and_object | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | para_crawl_enes | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/para_crawl_enes | lora | | ropes_background_situation_middle | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | adversarial_qa_dbert_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | ropes_prompt_beginning | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | cot_strategyqa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_strategyqa | lora | | duorc_ParaphraseRC_question_answering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | adversarial_qa_dbert_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_hop_original_generate_subject | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | race_high_Select_the_best_answer_generate_span_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | cot_gsm8k | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_gsm8k | lora | | glue_mrpc_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cos_e_v1_11_explain_why_human | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | race_high_Write_a_multi_choice_question_options_given_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | app_reviews_categorize_rating_using_review | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | adversarial_qa_dbert_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | quail_context_question_answer_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | qasc_qa_with_combined_facts_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | quoref_What_Is_The_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | glue_stsb_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | qasc_qa_with_separated_facts_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | web_questions_short_general_knowledge_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_testing_students | lora | | cnn_dailymail_3_4_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | kilt_tasks_hotpotqa_complex_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | huggingface_xsum | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/huggingface_xsum | lora | | cot_sensemaking_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | glue_sst2_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | dbpedia_14_pick_one_category_for_the_following_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | qasc_is_correct_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wiki_qa_Generate_Question_from_Topic | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_combining_facts | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quartz_given_the_fact_answer_the_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | wmt16_translate_ro_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | cot_sensemaking | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_sensemaking | lora | | wiki_bio_what_content | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | quail_context_question_description_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | quartz_answer_question_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | fix_punct | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/fix_punct | lora | | qasc_is_correct_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | cos_e_v1_11_question_description_option_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | wmt16_translate_fi_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | wiki_qa_Is_This_True_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | web_questions_question_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_question_answer | lora | | quartz_use_info_from_paragraph_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_choose_between | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_choose_between | lora | | adversarial_qa_droberta_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | cot_ecqa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_ecqa | lora | | web_questions_get_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_movie_director | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | dream_answer_to_dialogue | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | ropes_plain_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | cos_e_v1_11_rationale | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | duorc_ParaphraseRC_build_story_around_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | kilt_tasks_hotpotqa_straighforward_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | wiki_hop_original_explain_relation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | natural_questions_open_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | anli_r1_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | quoref_Given_Context_Answer_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | duorc_SelfRC_answer_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | wmt16_translate_de_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | duorc_ParaphraseRC_extract_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | unified_qa_science_inst | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | yelp_polarity_reviews_0_2_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | quarel_logic_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | wiki_qa_automatic_system | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | duorc_SelfRC_build_story_around_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | quartz_having_read_above_passage | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | glue_cola_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | wiqa_effect_with_string_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | duorc_ParaphraseRC_answer_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | ag_news_subset_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | dbpedia_14_given_a_choice_of_categories_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | dream_baseline | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_baseline | lora | | qasc_qa_with_separated_facts_4 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | race_middle_Write_a_multi_choice_question_options_given_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | qasc_qa_with_separated_facts_5 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | duorc_ParaphraseRC_decide_worth_it | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | duorc_SelfRC_question_answering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | definite_pronoun_resolution_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | super_glue_rte_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | stream_qed | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_qed | lora | | app_reviews_convert_to_rating | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiqa_what_might_be_the_first_step_of_the_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | social_i_qa_Show_choices_and_generate_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | cos_e_v1_11_generate_explanation_given_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | quartz_use_info_from_question_paragraph | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | anli_r2_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | duorc_ParaphraseRC_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quarel_heres_a_story | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | cos_e_v1_11_aligned_with_common_sense | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | duorc_SelfRC_extract_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | race_middle_Select_the_best_answer_no_instructions_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | web_questions_potential_correct_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | adversarial_qa_dbidaf_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_qasc | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_qasc | lora | | adversarial_qa_dbert_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | paws_wiki_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | quail_context_description_question_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | kilt_tasks_hotpotqa_final_exam | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | trec_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_effect_with_label_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | snli_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/snli_1_1_0 | lora | | cot_ecqa_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | quail_context_question_answer_description_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | gigaword_1_2_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | cos_e_v1_11_question_option_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | glue_qnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | duorc_SelfRC_generate_question_by_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | kilt_tasks_hotpotqa_formulate | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | quartz_paragraph_question_plain_concat | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | adversarial_qa_dbert_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | ropes_prompt_bottom_hint_beginning | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | adversarial_qa_droberta_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | dream_generate_first_utterance | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | duorc_SelfRC_decide_worth_it | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | quail_context_description_question_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | race_high_Is_this_the_right_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/drop_2_0_0 | lora | | ropes_read_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | glue_wnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | wiki_qa_Direct_Answer_to_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | web_questions_whats_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | quail_no_prompt_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | duorc_ParaphraseRC_generate_question_by_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | cos_e_v1_11_question_option_description_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | super_glue_copa_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | app_reviews_generate_review | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_hop_original_choose_best_object_affirmative_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | app_reviews_convert_to_star_rating | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | gem_web_nlg_en_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | quoref_Context_Contains_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | gem_e2e_nlg_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | quoref_Answer_Question_Given_Context | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | duorc_SelfRC_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | race_high_Select_the_best_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | stream_qed_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_qed_ii | lora | | cos_e_v1_11_description_question_option_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_bio_comprehension | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | duorc_ParaphraseRC_title_generation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_no_prompt_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | adversarial_qa_dbidaf_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | quoref_Found_Context_Online | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | race_middle_Taking_a_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | gem_common_gen_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | race_high_Taking_a_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | gem_dart_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | wiki_bio_guess_person | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quac_1_0_0 | lora | | cosmos_qa_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | lambada_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | ropes_given_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | adversarial_qa_droberta_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | adversarial_qa_dbidaf_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | adversarial_qa_dbidaf_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiki_hop_original_choose_best_object_affirmative_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | cot_esnli_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quail_description_context_question_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | wiki_bio_who | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_who | lora | Last updated on: 2024-03-15 17:41:32+00:00
[ "SCIQ" ]
SilasK/llama-7b-medqa_version_7
SilasK
null
[ "peft", "tensorboard", "safetensors", "llama", "trl", "sft", "generated_from_trainer", "base_model:huggyllama/llama-7b", "base_model:adapter:huggyllama/llama-7b", "license:other", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-13T12:49:38Z"
2024-03-13T23:17:13+00:00
0
0
--- base_model: huggyllama/llama-7b library_name: peft license: other tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-medqa_version_7 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-medqa_version_7 This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 1 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
ShivamChadha/whisper-medical-data
ShivamChadha
null
[ "peft", "tensorboard", "safetensors", "generated_from_trainer", "base_model:openai/whisper-small.en", "base_model:adapter:openai/whisper-small.en", "license:apache-2.0", "region:us" ]
"2024-03-14T16:45:57Z"
2024-03-14T16:45:59+00:00
0
0
--- base_model: openai/whisper-small.en library_name: peft license: apache-2.0 metrics: - wer tags: - generated_from_trainer model-index: - name: whisper-medical-data results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # whisper-medical-data This model is a fine-tuned version of [openai/whisper-small.en](https://huggingface.co/openai/whisper-small.en) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.5610 - Wer: 29.2608 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 50 - training_steps: 100 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 1.2874 | 16.67 | 100 | 1.5610 | 29.2608 | ### Framework versions - PEFT 0.9.1.dev0 - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
[ "MEDICAL DATA" ]
fakufaku/diffsep
fakufaku
null
[ "license:mit", "region:us" ]
"2024-03-15T07:29:07Z"
2024-03-15T07:37:59+00:00
0
0
--- license: mit --- Diffusion-based Generative Speech Source Separation This repository contains the checkpoints for the diffusion based speech separation model from the paper Diffusion-based Generative Speech Source Separation presented at ICASSP 2023. The code to run the model is available on [github](https://github.com/fakufaku/diffusion-separation). ### Abstract We propose DiffSep, a new single channel source separation method based on score-matching of a stochastic differential equation (SDE). We craft a tailored continuous time diffusion-mixing process starting from the separated sources and converging to a Gaussian distribution centered on their mixture. This formulation lets us apply the machinery of score-based generative modelling. First, we train a neural network to approximate the score function of the marginal probabilities or the diffusion-mixing process. Then, we use it to solve the reverse time SDE that progressively separates the sources starting from their mixture. We propose a modified training strategy to handle model mismatch and source permutation ambiguity. Experiments on the WSJ0 2mix dataset demonstrate the potential of the method. Furthermore, the method is also suitable for speech enhancement and shows performance competitive with prior work on the VoiceBank-DEMAND dataset. ID: `2022-10-23_01-37-07_experiment-model-large-multigpu_model.optimizer.lr-0.0002_model.sde.d_lambda-2.0_model.sde.sigma_min-0.05_epoch-979_si_sdr-11.271_N-30_snr-0.5_corrstep-1_denoise-True_schedule-None`
[ "CRAFT" ]
SilasK/llama-7b-medqa_version_9
SilasK
null
[ "peft", "tensorboard", "safetensors", "llama", "trl", "sft", "generated_from_trainer", "base_model:huggyllama/llama-7b", "base_model:adapter:huggyllama/llama-7b", "license:other", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-15T09:55:40Z"
2024-03-17T01:17:49+00:00
0
0
--- base_model: huggyllama/llama-7b library_name: peft license: other tags: - trl - sft - generated_from_trainer model-index: - name: llama-7b-medqa_version_9 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # llama-7b-medqa_version_9 This model is a fine-tuned version of [huggyllama/llama-7b](https://huggingface.co/huggyllama/llama-7b) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.2 - train_batch_size: 10 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 2 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
noguchis/viewdiff_train_teddybear
noguchis
null
[ "diffusers", "safetensors", "license:cc-by-nc-sa-4.0", "diffusers:CustomStableDiffusionPipeline", "region:us" ]
"2024-03-18T10:11:31Z"
2024-03-18T11:49:40+00:00
0
3
--- license: cc-by-nc-sa-4.0 --- This model I have uploaded is checkpoint 1061000 for Teddy Bear (CO3D) trained using [ViewDiff](https://github.com/facebookresearch/ViewDiff). Please see my article for details --> [Trying ViewDiff with WSL2...but](https://note.com/ngc_shj/n/n50f96770c0bf)
[ "BEAR" ]
mpetitguillaume/cryptoGPT-1.0-7B-lt
mpetitguillaume
text-generation
[ "transformers", "pytorch", "mistral", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
"2024-03-19T08:55:39Z"
2024-03-20T16:49:46+00:00
0
3
--- {} --- # CryptoGPT-1.0-7B-lt - Sentiment analysis model for financial news dedicated to crypto-assets This project introduces our lightest AI model designed for real-time market sentiment analysis on financial news dedicated to cryptoassets. Leveraging the latest advances in natural language processing (NLP), our model classifies financial texts into specific classes and evaluates their sentiment, providing invaluable insights into market trends. ### 1. Background and Problem Statement The cryptoasset market is known for its volatility, driven by various factors including statements by influential figures, political decisions, rumors, etc. Traditional financial models often fail to accurately interpret the impact of such events. In response to this challenge and inspired by the introduction of BloombergGPT, our project aims to democratize access to cutting-edge NLP for financial analysis dedicated to the crypto-asset market. Our research focused on developing a more accessible Large Language Model (LLM) capable of analyzing tweets, recent news, and market trends with limited resources. ### 2. Annotation Methodology ##### 2.1. Annotation Objective Our methodology is built around the annotation of financial texts into 21 financial classes specific to crypto-assets, and having a significant impact on this market, such as “Regulation and legislation”, “Market sentiment”, “ESG impact”, etc. The objective is twofold: to enable very precise categorization and to provide a comprehensive analysis of the sentiment reflected by the technical nuances of financial markets, particularly in the area of crypto-assets. ##### 2.2. Annotation Method We have used one of our automatic annotation technologies as part of our annotation and categorization process to ensure greater reliability compared to human annotation. Our input dataset of more than 15 million tokens has allowed us to develop high-performance market sentiment analysis models dedicated to crypto-assets. The cryptoGPT-1.0-7B-lt model that we present to you in this repository is fine-tuned on a much smaller input dataset of around 3.3 million tokens. #### 3. Fine-Tuning Strategy Our fine-tuning process uses broad language models with QLoRA for efficient adaptation. We have optimized the training phase of our models to run on a small infrastructure, ensuring significant resource and time savings without compromising model performance. ### 4. Installation To set up the environment for our model, follow these steps: ```bash !pip install --upgrade pip !pip install -q -U bitsandbytes !pip install -q -U git+https://github.com/huggingface/transformers.git !pip install -q -U git+https://github.com/huggingface/peft.git !pip install -q -U git+https://github.com/huggingface/accelerate.git !pip install -q huggingface_hub ``` Alternatively, you can create your requiremnts.txt and install all required packages like below: ```python bitsandbytes git+https://github.com/huggingface/transformers.git git+https://github.com/huggingface/peft.git git+https://github.com/huggingface/accelerate.git huggingface_hub tokenizers==0.15.2 ``` ```bash !pip install --upgrade pip !pip install -r requirements.txt ``` ### 5. Usage The model can be used to analyze financial texts ideally dedicated to the crypto-asset market, thereby providing accurate technical analysis of sentiment in various categories relevant to the crypto-asset market. ### 6. Python Example Code Here is a simple example of Python code to illustrate a basic use of the model for sentiment analysis: ```python import torch from transformers import GenerationConfig, AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig from huggingface_hub import login import gc gc.collect() torch.cuda.empty_cache() HF_TOKEN = "HF_TOKEN" login(HF_TOKEN) MODEL_NAME = "mpetitguillaume/cryptoGPT-1.0-7B-lt" def setup_device(): """Configures and returns the primary device for model computations (GPU if available).""" return torch.device("cuda") def login_to_hf_hub(token): """Authenticates with the Hugging Face Hub using a provided token.""" login(token=token) def load_model_and_tokenizer(model_name, bnb_config): """ Loads the specified model and tokenizer from Hugging Face, applying quantization configurations if provided. Also sets the tokenizer's pad token to its eos token. Args: model_name (str): Name of the model to load. bnb_config (BitsAndBytesConfig): Configuration for model quantization. Returns: tuple: The loaded model and tokenizer. """ model = AutoModelForCausalLM.from_pretrained( model_name, quantization_config=bnb_config, device_map="auto", trust_remote_code=True, ) tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) tokenizer.pad_token = tokenizer.eos_token # Harmonize pad and eos tokens return model, tokenizer def create_bnb_config(): """Creates a BitsAndBytes configuration optimized for model performance.""" return BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type="nf4", bnb_4bit_compute_dtype=torch.bfloat16, ) def create_prompt_formats(news): """ Creates a formatted prompt template for a prompt in the instruction dataset :param sample: Prompt or sample from the instruction dataset """ INTRO_BLURB = "Below is an instruction that describes a task. Write a response that appropriately completes the request." INSTRUCTION_KEY = "### Instruction:" INPUT_KEY = "### Input:" RESPONSE_KEY = "### Response:" prompt = """You need to relate news to some of the following 21 categories, provide a brief explanation, conduct sentiment analysis within each category, and offer an overall sentiment analysis, especially focusing on the financial markets, including cryptocurrencies. Determine whether it has a strongly positive, moderately positive, strongly negative, moderately negative, or negligible impact on market trends. Categories: 1. Regulation and Legislation (keywords: government, SEC, regulator, law, regulation, legislation.) 2. Adoption and Usage (keywords: adoption, usage, business, institution, partnership.) 3. Geopolitical Events (keywords: geopolitics, conflict, election, economic policy.) 4. Technology and Infrastructure (keywords: technology, infrastructure, updates, protocols, security.) 5. Financial Market Performances (keywords: stock market, bond market, currencies, indices.) 6. Market Sentiment (keywords: sentiment, confidence, opinion, investors.) 7. Competition Between Cryptocurrencies (keywords: competition, fork, updates, new projects, cryptocurrencies.) 8. Partnerships and Collaborations (keywords: partnership, collaboration, business, institution.) 9. Initial Coin Offerings (keywords: ICO, token sales, fundraising, crowdfunding.) 10. Media Coverage (keywords: media, media coverage, reporting, articles, news.) 11. Exchange Listings (keywords: exchange platforms, listing, liquidity.) 12. Exchange Delistings (keywords: exchange platforms, delisting, liquidity.) 13. Exchange Volume and Liquidity (keywords: volume, liquidity, exchange, trading.) 14. Market Manipulation and Fraud (keywords: manipulation, fraud, deception, investigation.) 15. Influential Players' Interventions (keywords: influence, statements, personalities, entrepreneurs, analysts.) 16. Expert Analysis and Forecasts (keywords: analysis, forecasts, experts, projections, predictions.) 17. Integration with Financial Services (keywords: integration, financial services, banking, payments.) 18. Macroeconomic Indicators (keywords: macroeconomics, inflation, interest rates, economic growth.) 19. Cryptocurrency Events and Conferences (keywords: events, conferences, summits, forums, exhibitions related to cryptocurrencies.) 20. Rumors and Speculations (keywords: rumors, speculations, buzz, leaks, unconfirmed information.) 21. Impact ESG (ESG Impact) (keywords: environment, social, governance, sustainability, responsibility, ethics, impact, carbon footprint, energy consumption, mining, electronic waste, working conditions, transparency, corporate governance, diversity, inclusion, human rights, climate change.) If you don't know the category, response "OTHERS".""" blurb = f"{INTRO_BLURB}" instruction = f"{INSTRUCTION_KEY}\n{prompt}" input_context = f"{INPUT_KEY}\n{news}" if news else None response = f"{RESPONSE_KEY}\n" parts = [part for part in [blurb, instruction, input_context, response] if part] formatted_prompt = "\n\n".join(parts) return formatted_prompt def generate_response(model, tokenizer, news): """ Generates a text response for a given input news snippet using the model and tokenizer. Args: model (AutoModelForCausalLM): The model for generating responses. tokenizer (AutoTokenizer): The tokenizer for processing input and output texts. news (str): The news snippet to respond to. Returns: str: The generated text response. """ generation_config = GenerationConfig( max_new_tokens=300, do_sample=True, top_p=0.1, temperature=0.01, pad_token_id=tokenizer.eos_token_id, ) input_tensor = tokenizer(create_prompt_formats(news), return_tensors="pt", truncation=True) device = setup_device() with torch.inference_mode(): outputs = model.generate( input_ids=input_tensor["input_ids"].to("cuda"), attention_mask=input_tensor["attention_mask"], generation_config=generation_config, ) result = tokenizer.batch_decode( outputs.detach().cpu().numpy(), skip_special_tokens=True ) return result[0].split('### Response:')[1].split('###')[0] bnb_config = create_bnb_config() model, tokenizer = load_model_and_tokenizer(MODEL_NAME, bnb_config) ``` Here is an example output for this code: ```bash 0. Financial Market Performances (mention of Jinzhou Bank's financial situation) Neutral - The article suggests that Jinzhou Bank's financial situation was in good shape, with low bad-debt level and only a small percentage of personal-business loans having gone sour. 1. Market Manipulation and Fraud (mention of the potential involvement of a billionaire, Li Hejun, in the bank's problems) Negative - The article suggests that the billionaire, Li Hejun, may be behind the bank's distress, indicating potential market manipulation or fraud. 2. Geopolitical Events (mention of China's banks and its economic policies) Neutral - The article discusses China's banks and their financial situation, but it doesn't provide any clear geopolitical analysis or opinion. Sentiment Analysis regarding the Cryptocurrency Market: Somewhat Negative Impact on the Market: The article suggests that Jinzhou Bank's financial situation may be in trouble, and the billionaire Li Hejun might be involved. This could lead to rumors and speculations about the stability of China's banks and the economy. However, the article doesn't provide any clear sentiment analysis or opinion on the matter... ``` ### 7. Evaluation Results Our model evaluation was based on manual expert evaluation. As part of the evaluation of this very lightweight model, we selected a set of 50 financial articles dedicated to crypto-assets, representative of various categories, rich in content and representative of several market trends. Six models were tested: refined LLaMa-2-7B, refined Mistral-7B, LLaMa-2-7B, LLaMa-2-13B, Mistral-7B, GPT-3.5 Turbo. Each model was rated on a scale of 0 to 4, where 4 indicates optimal performance and 0 means unusable results. According to the data our model based on fine-tuned Mistral-7B showed superior performance to GPT-3.5. Our tests on our larger models demonstrated performance well above GPT-4 and the largest known wide language models. | Models | GPT-3.5 | **CryptoGPT-1.0-7B-lt** | Mistral-7B | LLaMa-2-7B | LLaMa-2-13B | |--- |:-: |:-: |:-: |--: |--: | | Average score | 2.9 | **3.12** | 0.48 | 0.38 | 0.68 | | Score 4 | 14 | **15** | 0 | 0 | 0 | | Score 4 & 3 | 35 | **41** | 0 | 0 | 0 | ### 8. Reporting Issues Please report any software "bug," or other problems with the models through one of the following means: - Reporting issues with the model: [[email protected]](mailto:[email protected]) ### 9. License This project is licensed under the MIT License - see the LICENSE file for details. ### 10. Contact For any questions or to contribute to the project, please contact us at [[email protected]](mailto:[email protected]).
[ "BLURB" ]
SilasK/mistral-7b-medqa-version4
SilasK
null
[ "peft", "tensorboard", "safetensors", "mistral", "trl", "sft", "generated_from_trainer", "base_model:mistralai/Mistral-7B-Instruct-v0.1", "base_model:adapter:mistralai/Mistral-7B-Instruct-v0.1", "license:apache-2.0", "4-bit", "bitsandbytes", "region:us" ]
"2024-03-19T10:54:21Z"
2024-03-19T16:21:01+00:00
0
0
--- base_model: mistralai/Mistral-7B-Instruct-v0.1 library_name: peft license: apache-2.0 tags: - trl - sft - generated_from_trainer model-index: - name: mistral-7b-medqa-version4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-medqa-version4 This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 20 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 1 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.37.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.16.1 - Tokenizers 0.15.0
[ "MEDQA" ]
WeMake/V41
WeMake
any-to-any
[ "not-for-all-audiences", "any-to-any", "en", "dataset:WeMake/Intelligent-Content-Understanding", "region:us" ]
"2024-03-20T03:42:16Z"
2025-03-01T14:51:14+00:00
0
1
--- datasets: - WeMake/Intelligent-Content-Understanding language: - en pipeline_tag: any-to-any tags: - not-for-all-audiences --- # V41 **V41** is WeMake's flagship multimodal AI model, designed to bridge the gap between different data formats and modalities. As an any-to-any model, V41 excels at tasks requiring the transformation, understanding, and generation of content across various formats including text, images, audio, and structured data. Built on WeMake's proprietary Intelligent Content Understanding (ICU) methodology, V41 represents a significant leap forward in creating AI that truly understands context, emotional nuances, and complex relationships between different types of information. ## Key Features - **Multimodal Versatility:** Seamlessly processes and generates across text, image, audio, and structured data formats - **Emotional Intelligence:** Incorporates advanced emotional understanding capabilities, enabling more human-like interactions - **Context-Aware Processing:** Maintains context across different modalities and extended interactions - **Ethical Alignment:** Designed with WeMake's commitment to ethical AI, prioritizing human values and well-being - **Adaptive Learning:** Continuously improves through our AI-Human Alignment Process - **Enterprise-Grade Security:** Implements robust security measures in compliance with European data protection standards ## Applications V41 powers a wide range of applications across industries: - **Productivity Enhancement:** Core of the V41 Platform with Clarity and Val Intelligence Workers - **Content Creation:** Generates cohesive multi-format content for marketing, education, and entertainment - **Data Analysis:** Transforms complex datasets into accessible visualizations and insights - **Customer Experience:** Powers conversational interfaces that understand emotional context and user needs - **Healthcare Support:** Assists in medical data interpretation and patient communication - **Educational Tools:** Creates interactive learning experiences across multiple formats ## Technical Architecture V41 is built on a transformer-based architecture with specialized encoders and decoders for each supported modality. The model features: - **Advanced Cross-Attention Mechanisms:** Enabling seamless information flow between modalities - **Contextual Memory System:** Maintaining coherence across extended interactions - **Ethical Guardrails:** Built-in parameters to ensure outputs align with WeMake's ethical principles - **Optimized Performance:** Balancing computational efficiency with high-quality outputs ## Model Variants V41 comes in several specialized variants: - **V41 Core:** The foundation model supporting all modalities - **Llama-3-8B-V41-Instruct-1048k:** Text-optimized version with 1M token context length - **VX-Unholy-13B:** Specialized creative variant built on Unholy-v2-13B - **V41-Vision:** Image-focused variant with enhanced visual understanding ## Responsible Use V41 is designed for beneficial applications that enhance human capabilities while respecting privacy and ethical boundaries. Users are expected to adhere to WeMake's ethics policy and applicable regulations when deploying V41-based applications. ## Training Methodology V41 was trained using WeMake's proprietary ICU (Intelligent Content Understanding) dataset and methodology, which focuses on building an internal knowledge map that enables the model to make connections between concepts across different modalities. This training approach emphasizes: - **Deep Contextual Understanding:** Beyond surface-level pattern recognition - **Ethical Considerations:** Minimizing harmful biases and promoting inclusive outputs - **Emotional Intelligence:** Recognizing and appropriately responding to emotional cues - **Human-AI Alignment:** Ensuring the model's outputs align with human values and expectations ## License V41 is released under a custom license that allows for research and commercial applications while prohibiting harmful or unethical use cases. Please contact WeMake for specific licensing details. ## Acknowledgments V41 builds upon the collective efforts of the AI research community while incorporating WeMake's innovations in emotional intelligence and ethical AI development. We extend our gratitude to all contributors to the field whose work has made V41 possible. ## Contact For inquiries about V41, please contact us at [[email protected]](mailto:[email protected]). Join us in building a future where AI amplifies human potential while respecting our values and well-being. 💙 **The WeMake Team** [🤝 Ethics Policy](https://wemake.cx/legal/ethics/) [🛡️ Privacy Policy](https://wemake.cx/legal/privacy) [📇 Imprint](https://wemake.cx/legal/imprint)
[ "MEDICAL DATA" ]
ostapeno/library-stablelm_flan_5ep_part2
ostapeno
null
[ "region:us" ]
"2024-03-20T15:19:52Z"
2024-03-22T21:23:53+00:00
0
0
--- {} --- Number of experts present in the library: 156 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | aeslc_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | race_middle_Is_this_the_right_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | true_case | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/true_case | lora | | quarel_do_not_use | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quarel_do_not_use | lora | | word_segment | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | stream_aqua_ii | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/stream_aqua_ii | lora | | wiki_hop_original_choose_best_object_affirmative_3 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | social_i_qa_Show_choices_and_generate_index | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | quail_context_question_description_answer_id | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | ropes_background_new_situation_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | wiki_bio_who | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_bio_who | lora | | gem_wiki_lingua_english_en_1_1_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | wiki_qa_Topic_Prediction_Answer_Only | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | trivia_qa_rc_1_1_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | wiqa_what_is_the_final_step_of_the_following_process | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | quail_context_question_description_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | wiki_hop_original_generate_subject_and_object | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | ropes_background_situation_middle | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | ropes_prompt_beginning | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | wiki_hop_original_generate_subject | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | race_high_Select_the_best_answer_generate_span_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | race_high_Write_a_multi_choice_question_options_given_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | quail_context_question_answer_description_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | quoref_What_Is_The_Answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | web_questions_short_general_knowledge_q | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quarel_testing_students | lora | | cnn_dailymail_3_4_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | huggingface_xsum | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/huggingface_xsum | lora | | wiki_qa_Generate_Question_from_Topic | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | quartz_given_the_fact_answer_the_q | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | wmt16_translate_ro_en_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | wiki_bio_what_content | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | quail_context_question_description_answer_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | quartz_answer_question_based_on | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | wmt16_translate_fi_en_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | wiki_qa_Is_This_True_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | web_questions_question_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/web_questions_question_answer | lora | | quartz_use_info_from_paragraph_question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_choose_between | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quarel_choose_between | lora | | web_questions_get_the_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | dream_answer_to_dialogue | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | ropes_plain_background_situation | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | duorc_ParaphraseRC_build_story_around_qa | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | wiki_hop_original_explain_relation | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | quoref_Given_Context_Answer_Question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | wmt16_translate_de_en_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | unified_qa_science_inst | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | yelp_polarity_reviews_0_2_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | quarel_logic_test | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | wiki_qa_automatic_system | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | duorc_SelfRC_build_story_around_qa | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | quartz_having_read_above_passage | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | wiqa_effect_with_string_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | qasc_qa_with_separated_facts_4 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | race_middle_Write_a_multi_choice_question_options_given_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | qasc_qa_with_separated_facts_5 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | super_glue_rte_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_what_might_be_the_first_step_of_the_process | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | social_i_qa_Show_choices_and_generate_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | quartz_use_info_from_question_paragraph | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | quarel_heres_a_story | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | race_middle_Select_the_best_answer_no_instructions_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | web_questions_potential_correct_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | quail_context_description_question_answer_id | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | trec_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_effect_with_label_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | snli_1_1_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/snli_1_1_0 | lora | | quail_context_question_answer_description_id | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | quartz_paragraph_question_plain_concat | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | ropes_prompt_bottom_hint_beginning | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | quail_context_description_question_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | race_high_Is_this_the_right_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/drop_2_0_0 | lora | | ropes_read_background_situation | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | wiki_qa_Direct_Answer_to_Question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | web_questions_whats_the_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | quail_no_prompt_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | super_glue_copa_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | wiki_hop_original_choose_best_object_affirmative_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | quoref_Context_Contains_Answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | quoref_Answer_Question_Given_Context | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | race_high_Select_the_best_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | stream_qed_ii | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/stream_qed_ii | lora | | wiki_bio_comprehension | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | quail_no_prompt_id | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | quoref_Found_Context_Online | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | race_middle_Taking_a_test | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | race_high_Taking_a_test | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | wiki_bio_guess_person | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quac_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | ropes_given_background_situation | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | wiki_hop_original_choose_best_object_affirmative_1 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | quail_description_context_question_answer_text | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | stream_qed | stabilityai/stablelm-3b-4e1t | sordonia/flan-10k-flat/stream_qed | lora | Last updated on: 2024-03-22 21:23:52+00:00
[ "SCIQ" ]
ostapeno/library-mistral7B_flan_5ep_higher_lr
ostapeno
null
[ "region:us" ]
"2024-03-23T19:08:25Z"
2024-04-05T18:13:32+00:00
0
0
--- {} --- Number of experts present in the library: 256 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | aeslc_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | race_middle_Is_this_the_right_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | true_case | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/true_case | lora | | cot_strategyqa_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | quarel_do_not_use | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_do_not_use | lora | | word_segment | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/word_segment | lora | | sciq_Multiple_Choice_Question_First | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | wiki_qa_Decide_good_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | stream_aqua_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_aqua_ii | lora | | wiki_hop_original_choose_best_object_affirmative_3 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | sciq_Direct_Question_Closed_Book_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | social_i_qa_Generate_the_question_from_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | duorc_ParaphraseRC_movie_director | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | math_dataset_algebra__linear_1d_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cot_gsm8k_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | social_i_qa_Show_choices_and_generate_index | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | wiqa_what_is_the_missing_first_step | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | dream_generate_last_utterance | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | quail_context_question_description_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | cot_creak_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_creak_ii | lora | | ropes_background_new_situation_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | quoref_Guess_Title_For_Context | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | imdb_reviews_plain_text_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | cot_esnli | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_esnli | lora | | anli_r3_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | adversarial_qa_droberta_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | wiki_bio_who | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_who | lora | | cos_e_v1_11_i_think | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | gem_wiki_lingua_english_en_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | ropes_prompt_bottom_no_hint | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | glue_qqp_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_description_question_option_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | wiki_qa_Topic_Prediction_Answer_Only | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | cot_creak | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_creak | lora | | trivia_qa_rc_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | duorc_SelfRC_title_generation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | wiqa_what_is_the_final_step_of_the_following_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | glue_mnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | quail_context_question_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | quoref_Guess_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | adversarial_qa_dbidaf_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | adversarial_qa_droberta_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | cos_e_v1_11_question_description_option_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | wiki_hop_original_generate_subject_and_object | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | para_crawl_enes | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/para_crawl_enes | lora | | ropes_background_situation_middle | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | adversarial_qa_dbert_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | ropes_prompt_beginning | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | cot_strategyqa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_strategyqa | lora | | duorc_ParaphraseRC_question_answering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | adversarial_qa_dbert_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | wiki_hop_original_generate_subject | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | race_high_Select_the_best_answer_generate_span_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | cot_gsm8k | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_gsm8k | lora | | glue_mrpc_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | cos_e_v1_11_explain_why_human | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | race_high_Write_a_multi_choice_question_options_given_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | app_reviews_categorize_rating_using_review | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | adversarial_qa_dbert_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | quail_context_question_answer_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | qasc_qa_with_combined_facts_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | quoref_What_Is_The_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | quail_context_description_question_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | glue_stsb_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | qasc_qa_with_separated_facts_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | web_questions_short_general_knowledge_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | quarel_testing_students | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_testing_students | lora | | cnn_dailymail_3_4_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | wiki_bio_key_content | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | sciq_Multiple_Choice | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | kilt_tasks_hotpotqa_complex_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | huggingface_xsum | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/huggingface_xsum | lora | | cot_sensemaking_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | glue_sst2_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | dbpedia_14_pick_one_category_for_the_following_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | qasc_is_correct_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wiki_qa_Generate_Question_from_Topic | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_combining_facts | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | quartz_given_the_fact_answer_the_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | wmt16_translate_ro_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | cot_sensemaking | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_sensemaking | lora | | wiki_bio_what_content | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | sciq_Direct_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | quail_context_question_description_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | ropes_prompt_mix | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | quartz_answer_question_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | fix_punct | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/fix_punct | lora | | qasc_is_correct_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | cos_e_v1_11_question_description_option_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | wmt16_translate_fi_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | super_glue_multirc_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | wiki_qa_Is_This_True_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | multi_news_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | web_questions_question_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_question_answer | lora | | quartz_use_info_from_paragraph_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wmt14_translate_fr_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | quarel_choose_between | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_choose_between | lora | | adversarial_qa_droberta_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | cot_ecqa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_ecqa | lora | | web_questions_get_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_movie_director | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | dream_answer_to_dialogue | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | ropes_plain_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | cos_e_v1_11_rationale | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | duorc_ParaphraseRC_build_story_around_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | social_i_qa_I_was_wondering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_qa_exercise | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | kilt_tasks_hotpotqa_straighforward_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | wiki_hop_original_explain_relation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | wiki_hop_original_choose_best_object_interrogative_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | natural_questions_open_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | anli_r1_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | quoref_Given_Context_Answer_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | duorc_SelfRC_answer_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | wmt16_translate_de_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | wiki_hop_original_generate_object | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | duorc_ParaphraseRC_extract_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | unified_qa_science_inst | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quail_description_context_question_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | quarel_logic_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_logic_test | lora | | squad_v1_1_3_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | wiki_qa_Jeopardy_style | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | wiki_qa_automatic_system | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | ropes_new_situation_background_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | super_glue_wsc_fixed_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | wmt16_translate_tr_en_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | duorc_SelfRC_build_story_around_qa | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | quoref_Answer_Friend_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | quartz_having_read_above_passage | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | glue_cola_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | wiqa_effect_with_string_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | duorc_ParaphraseRC_answer_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | ag_news_subset_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | dbpedia_14_given_a_choice_of_categories_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | dream_baseline | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_baseline | lora | | qasc_qa_with_separated_facts_4 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | quartz_read_passage_below_choose | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | race_middle_Write_a_multi_choice_question_options_given_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | ropes_plain_no_background | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | qasc_qa_with_separated_facts_5 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | quoref_Read_And_Extract_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | duorc_ParaphraseRC_decide_worth_it | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | duorc_SelfRC_question_answering | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | definite_pronoun_resolution_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | super_glue_rte_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | social_i_qa_Generate_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | stream_qed | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_qed | lora | | app_reviews_convert_to_rating | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiqa_what_might_be_the_first_step_of_the_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | social_i_qa_Show_choices_and_generate_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | cos_e_v1_11_generate_explanation_given_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | quartz_use_info_from_question_paragraph | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | anli_r2_0_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | duorc_ParaphraseRC_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | quarel_heres_a_story | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | cos_e_v1_11_aligned_with_common_sense | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | duorc_SelfRC_extract_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | race_middle_Select_the_best_answer_no_instructions_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | web_questions_potential_correct_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | qasc_qa_with_separated_facts_3 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | adversarial_qa_dbidaf_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_qasc | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_qasc | lora | | adversarial_qa_dbert_tell_what_it_is | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | paws_wiki_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | quail_context_description_question_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | coqa_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | kilt_tasks_hotpotqa_final_exam | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | trec_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/trec_1_0_0 | lora | | super_glue_cb_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | qasc_qa_with_separated_facts_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | wiqa_effect_with_label_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | snli_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/snli_1_1_0 | lora | | cot_ecqa_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | quail_context_question_answer_description_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | gigaword_1_2_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | cos_e_v1_11_question_option_description_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | glue_qnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | duorc_SelfRC_generate_question_by_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | kilt_tasks_hotpotqa_formulate | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | quartz_paragraph_question_plain_concat | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | adversarial_qa_dbert_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | ropes_prompt_bottom_hint_beginning | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | adversarial_qa_droberta_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | dream_generate_first_utterance | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | duorc_SelfRC_decide_worth_it | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | quail_context_description_question_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | race_high_Is_this_the_right_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | drop_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/drop_2_0_0 | lora | | ropes_read_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | glue_wnli_2_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | wiki_qa_Direct_Answer_to_Question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | web_questions_whats_the_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiki_qa_found_on_google | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | quail_no_prompt_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | duorc_ParaphraseRC_generate_question_by_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | cos_e_v1_11_question_option_description_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | super_glue_copa_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | app_reviews_generate_review | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | wiki_hop_original_choose_best_object_affirmative_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | app_reviews_convert_to_star_rating | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | gem_web_nlg_en_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | quoref_Context_Contains_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | gem_e2e_nlg_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | quoref_Answer_Question_Given_Context | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quartz_answer_question_below | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | duorc_SelfRC_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | race_high_Select_the_best_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | stream_qed_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_qed_ii | lora | | cos_e_v1_11_description_question_option_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | wiki_bio_comprehension | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | duorc_ParaphraseRC_title_generation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_no_prompt_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | adversarial_qa_dbidaf_answer_the_following_q | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | quoref_Found_Context_Online | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | race_middle_Select_the_best_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | race_middle_Taking_a_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | quoref_Answer_Test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | gem_common_gen_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | race_high_Taking_a_test | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | quail_description_context_question_answer_id | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | gem_dart_1_1_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | wiki_bio_guess_person | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | race_high_Select_the_best_answer_no_instructions_ | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | quac_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quac_1_0_0 | lora | | cosmos_qa_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | dream_read_the_following_conversation_and_answer_the_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | wiki_qa_Topic_Prediction_Question_Only | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | lambada_1_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | ropes_given_background_situation | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | adversarial_qa_droberta_generate_question | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | adversarial_qa_dbidaf_question_context_answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | adversarial_qa_dbidaf_based_on | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | wiki_hop_original_choose_best_object_affirmative_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | super_glue_wic_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | ropes_plain_bottom_hint | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | quoref_Find_Answer | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | wiki_hop_original_choose_best_object_interrogative_1 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | cot_esnli_ii | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quail_description_context_question_answer_text | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | squad_v2_0_3_0_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | stream_aqua | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/stream_aqua | lora | | super_glue_record_1_0_2 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | yelp_polarity_reviews_0_2_0 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | Last updated on: 2024-03-25 21:05:58+00:00
[ "SCIQ" ]
ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-polarity
ginkgogo
text-classification
[ "setfit", "safetensors", "mpnet", "absa", "sentence-transformers", "text-classification", "generated_from_setfit_trainer", "arxiv:2209.11055", "base_model:sentence-transformers/all-mpnet-base-v2", "base_model:finetune:sentence-transformers/all-mpnet-base-v2", "model-index", "region:us" ]
"2024-03-23T23:19:03Z"
2024-03-24T23:47:26+00:00
0
0
--- base_model: sentence-transformers/all-mpnet-base-v2 library_name: setfit metrics: - accuracy pipeline_tag: text-classification tags: - setfit - absa - sentence-transformers - text-classification - generated_from_setfit_trainer widget: - text: ordered, great atmosphere, excellent service:FINALLY tried Mizza and wasn't disappointed. Loved (almost) everything we ordered, great atmosphere, excellent service, and the perfect setting for a lovely bday Sunday. The burrata & heirloom tomatoes app was scrumptious, the salmon pasta, very flavorful and the salmon perfectly cooked, I liked the toppings of the veggie pizza but wasn't a super fan of the crust (doesn't mean I won't come back and try another pizza on their menu ) and the cannoli was good although that dessert in general isn't my fave (it was my bf's bday so had to get what he wanted ). The flourless chocolate cake and limoncello cake are what I'll try next time. Had a great time and will be back. Gave it 4 stars just cuz I wasn't that excited about the pizza and that's something they're supposed to so well. Would recommend the restaurant though! - text: "is because the food was decent,:Three reasons why it gets three stars:\n\n\ 1. The crab cakes were good and is a definitely must try!\n2. The shrimp scampi\ \ was actually amazing in the sauce that it comes with, so that's another must\ \ try!\n3. The real reason why it is getting three stars is because service is\ \ everything in ANY restaurant you go to. Service started off great, waitress\ \ was attentive, but once we paid the bill and left a 20% tip, my guests and I,\ \ which was only three of us, stayed at the table to finish our drinks and we're\ \ looking at funny videos from a trip we went to. Point is the waitress rudely\ \ told my friend to lower the volume on his phone, yet other guests were just\ \ as loud and we were sitting OUTSIDE...where it is already a loud environment!\ \ \n\nI really want to give it 4 stars, but if I give 4 stars it changes it to,\ \ \"Yay! I'm a fan\", but I am not. The only reason why it's not getting 1 star,\ \ is because the food was decent, the view is nice and also the manager was extremely\ \ empathetic to the situation and it wasn't her fault at all that her waitress\ \ was obviously having an off day. I have never met a manager that attentive and\ \ she was incredible at handling and diffusing the situation. I cannot thank her\ \ enough for salvaging the rest of our evening for how poor the waitress treated\ \ paying customers." - text: and the perfect setting for a lovely:FINALLY tried Mizza and wasn't disappointed. Loved (almost) everything we ordered, great atmosphere, excellent service, and the perfect setting for a lovely bday Sunday. The burrata & heirloom tomatoes app was scrumptious, the salmon pasta, very flavorful and the salmon perfectly cooked, I liked the toppings of the veggie pizza but wasn't a super fan of the crust (doesn't mean I won't come back and try another pizza on their menu ) and the cannoli was good although that dessert in general isn't my fave (it was my bf's bday so had to get what he wanted ). The flourless chocolate cake and limoncello cake are what I'll try next time. Had a great time and will be back. Gave it 4 stars just cuz I wasn't that excited about the pizza and that's something they're supposed to so well. Would recommend the restaurant though! - text: ) and the service is friendly and:I'm not sure what what I would do if I'd never discovered Nikka, since it's the definitely the most authentic ramen one can get in the area. Prices are standard for ramen (especially in SB) and the service is friendly and efficient. Not only is Nikka's ramen amazing, their variety of appetizers is also great. I've yet to try one that I don't like. Definitely come here if you're looking to satisfy your ramen craving! - text: Overall an excellent experience and the friendly:I got a to-go order for empanadas on the lunch menu and it was fantastic. The dish was incredibly flavorful and the Kombucha the owner recommended was amazing. Overall an excellent experience and the friendly owner, waiters, and waitresses are just the cherry on top. I would highly recommend any vegetarians to try out this spot! inference: false model-index: - name: SetFit Polarity Model with sentence-transformers/all-mpnet-base-v2 results: - task: type: text-classification name: Text Classification dataset: name: Unknown type: unknown split: test metrics: - type: accuracy value: 0.65 name: Accuracy --- # SetFit Polarity Model with sentence-transformers/all-mpnet-base-v2 This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Aspect Based Sentiment Analysis (ABSA). This SetFit model uses [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification. In particular, this model is in charge of classifying aspect polarities. The model has been trained using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. This model was trained within the context of a larger system for ABSA, which looks like so: 1. Use a spaCy model to select possible aspect span candidates. 2. Use a SetFit model to filter these possible aspect span candidates. 3. **Use this SetFit model to classify the filtered aspect span candidates.** ## Model Details ### Model Description - **Model Type:** SetFit - **Sentence Transformer body:** [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance - **spaCy Model:** en_core_web_sm - **SetFitABSA Aspect Model:** [ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-aspect](https://huggingface.co/ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-aspect) - **SetFitABSA Polarity Model:** [ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-polarity](https://huggingface.co/ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-polarity) - **Maximum Sequence Length:** 384 tokens - **Number of Classes:** 4 classes <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit) - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055) - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit) ### Model Labels | Label | Examples | |:---------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | negative | <ul><li>'into more American food, added burgers:They made it into more American food, added burgers and ribs and got rid of the tequila selection. We were so bummed. Used to be one of our favorite places to go for good Mexican food. The owner said the new direction was to appeal to more tourists.'</li></ul> | | positive | <ul><li>"was great and seating was comfortable.:Such a cute little spot for desserts! I'm so glad we had time on our short visit to Santa Barbara to grab a slice of cake from here. My husband and I each got our own to slice to share of course. He said we didn't come all this way just to get one so we chose a slice of the berry cake and chocolate decadence. The berry cake was nice and fluffy without being too sweet. The acidity from the fruits balanced the sweetest of the cake wonderfully. If you're up for something rich then the chocolate decadence will not disappoint. Service was great and seating was comfortable. Order your sweet treats at the counter then a number will be given to you. Pick a table and get ready to enjoy because your sweets will be brought out to your table when ready."</li><li>'Soon, our food was delivered,:One brisk Saturday morning after asking workers during a stop for tylenol from the Hotel California Boutique the best breakfast place, they recommended Goat Tree. We crossed the busy street and greeted the hostess. The very kind young lady walked us to our table on the sunny patio. We skimmed the menu and decided on the chicken and waffle and a chocolate croissant. The wait was quite short and we spent it discussing the beautiful surrounding area. Soon, our food was delivered, and let me tell you, it was beautiful. On top of that, it was scrumptious. The fried chicken was perfect and tender. The waffle had the perfect balance of crunch and fluff. And how dare I forget the exquisite honey. Now this honey was the best I have ever tasted. It was topped with chia and pumpkin seeds. My daughter asked for her croissant warmed, and once again it was marvelous. After paying, I told our waitress how amazing the honey was. Next thing we knew, she brought out two large to go cups full of it! \n\nAbsolutely loved this place and everything about it. 100% recommend! I strongly award them 5 stars!'</li><li>". \n\nThe service was impeccable,:Man! I was so drunk when I ate here last weekend. \n\nI came up from LA to celebrate my boyfriend's best friend's graduation. So after the commencement ceremony a bunch of us went to a friend's house and had Vodka tonics. He made them great and I was drunk. So we went to dinner at this beautiful hotel overlooking the beach. \n\nThe best friend's parents bought us (about 20 people) dinner at Rodney's Steakhouse. There was a pre-set menu with different choices for us. \n\nFor the appetizer, I ordered the Sea Scallops. These were the best damn sea scallops I've ever had. They literally melted in my mouth and were so delicious. They came in a white wine and garlic butter onion tartlet with truffle vinaigrette.\n\nPlease keep in mind that the wine kept flowing and continued to get very giggly and drunk. So fun!\n\nFor the main course, I ordered the Roasted Dover Sole Fillet which had lump crab meat stuffing and lemon butter. It was good but it wasn't great. \n\nFor dessert I had the creme brulee which was strange tasting. I would have much rather had the chocolate mousse. \n\nThe service was impeccable, the bathrooms were very nice and clean and I met a lot of great people. Or so I think so. =)"</li></ul> | | mixed | <ul><li>"is because the food was inedible.:I rarely ever give anything less than a 2 star and if I do, it is because the food was inedible. Literally, we paid so much for our entrees and tried to force ourselves to eat it because we hate to waste food but we couldn't even do that. Maybe we ordered the wrong dish. We had the ramen and the risotto- and we've had these type of dishes many times before. In fact, it's one of our favorite dishes normally. But the ramen was sooo disappointing. It was just watered down soy sauce. Imagine how salty that is. I've never had anything like this and was completely shocked. And the risotto was so undercooked. I am okay with al dente but I am saying this was borderline raw, hard, and was hard to digest. BUT- the BONE MARROW was AMAZING. Do order this. And maybe only this. It was perfectly balanced and savory and aromatic and presented beautifully. That was the only good tapas that we ordered and why I will give it a 2 star instead of a 1 star. Also, we spoke to the manager about our food and we discounted us and was really nice about it. I felt pretty bad but I also thought they should know- maybe we just came on a bad day? I am just really surprised that this place had such a high rating. I took my mom here for her birthday and we left this restaurant hungry and disappointed"</li></ul> | | neutral | <ul><li>"limited amount of seating for the long:If you're ever missing LA street tacos, Lilly's is the closest you're gonna get. Without taking that into consideration, Lilly's is without a doubt one of the cheapest places to get a filling and delicious meal along the Pacific Coast, and you will love it.\n\nThe $1.80 tacos come out before you're even done ordering, which is wild considering that while they have a bunch of meats already prepped, there's still a constant rotation of beef, pork, and chicken on the grill. All tacos come double-wrapped, and if you eat in, on a Styrofoam plate that might betray the county's love for eco-friendly packaging, but it certainly cuts down on the costs.\n\nThe salsa bar is never consistent; while all the fixings are always well-stocked, how spicy each salsa is changes from day to day. Sometimes, it'll be the dark brown one that'll cause you to start crying; other days, it's the green one that packs a punch. Taste test each one before you squirt it on. Old Yelp reviews suggest their grilled onions and jalapenos used to be free, but it costs $1 for a small plate of them nowadays.\n\nBeing next to the 101 isn't ideal, nor is the limited amount of seating for the long line that builds up outside the store. But these are the kinds of things you get saddled with, not quite the stuff you can choose or imagine happening when you first start out.\n\nYou'll incur a $0.50 charge for credit cards if you spend less than $5 here. You're going to want so many tacos that you won't even think about it."</li></ul> | ## Evaluation ### Metrics | Label | Accuracy | |:--------|:---------| | **all** | 0.65 | ## Uses ### Direct Use for Inference First install the SetFit library: ```bash pip install setfit ``` Then you can load this model and run inference. ```python from setfit import AbsaModel # Download from the 🤗 Hub model = AbsaModel.from_pretrained( "ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-aspect", "ginkgogo/setfit-absa-bge-small-en-v1.5-restaurants-polarity", ) # Run inference preds = model("The food was great, but the venue is just way too busy.") ``` <!-- ### Downstream Use *List how someone could finetune this model on their own dataset.* --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Set Metrics | Training set | Min | Median | Max | |:-------------|:----|:---------|:----| | Word count | 25 | 123.9048 | 272 | | Label | Training Sample Count | |:---------|:----------------------| | mixed | 1 | | negative | 1 | | neutral | 1 | | positive | 18 | ### Training Hyperparameters - batch_size: (50, 50) - num_epochs: (5, 5) - max_steps: -1 - sampling_strategy: oversampling - body_learning_rate: (2e-05, 1e-05) - head_learning_rate: 0.01 - loss: CosineSimilarityLoss - distance_metric: cosine_distance - margin: 0.25 - end_to_end: False - use_amp: True - warmup_proportion: 0.1 - seed: 42 - eval_max_steps: -1 - load_best_model_at_end: True ### Training Results | Epoch | Step | Training Loss | Validation Loss | |:------:|:----:|:-------------:|:---------------:| | 0.1429 | 1 | 0.2034 | - | ### Framework Versions - Python: 3.10.12 - SetFit: 1.0.3 - Sentence Transformers: 2.6.0 - spaCy: 3.7.4 - Transformers: 4.38.2 - PyTorch: 2.2.1+cu121 - Datasets: 2.18.0 - Tokenizers: 0.15.2 ## Citation ### BibTeX ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
[ "CHIA" ]
dubious-bear/test
dubious-bear
null
[ "en", "region:us" ]
"2024-03-24T02:29:56Z"
2024-03-24T03:19:03+00:00
0
0
--- language: - en extra_gated_heading: You need to share contact information with ... to access this model extra_gated_prompt: ' ### DUBIOUS BEAR LICENSE "Agreement" means the terms and conditions for use, reproduction, distribution and modification of the Dubious Bear materials set forth herein. "Documentation" means the specifications, manuals and documentation accompanying ... #### Acceptable use Policy You agree you will not use, or allow others to use, ... Please report any violation of this Policy, software “bug,” or other problems that could lead to a violation of this Policy through one of the following means: * Reporting issues with the model: Link here * Reporting risky content generated by the model: Link here * Reporting bugs and security concerns: Link here * Reporting violations of the Acceptable Use Policy or unlicensed uses: Link here' extra_gated_fields: First Name: text Last Name: text Organization: text ? By clicking 'Submit' below, I accept the terms of the license and acknowledge that the information I provide will be collected, stored, processed, and shared in accordance with XXX : checkbox extra_gated_description: The information you provide will be collected, stored, processed and shared in accordance with the ... Privacy Policy. extra_gated_button_content: Submit inference: false ---
[ "BEAR" ]
sohampatil/msphi2-medical
sohampatil
null
[ "license:mit", "region:us" ]
"2024-03-24T15:04:44Z"
2024-03-24T15:16:39+00:00
0
3
--- license: mit --- --- library_name: peft base_model: microsoft/phi-2 --- ## Model Details ### Model Description We have developed a finetuned AI model designed to revolutionize how medical reports are generated. This model, built upon the robust foundation of Microsoft's Phi Large Language Model (LLM), leverages the specialized MedMCQA dataset, sourced from the OpenLifeSciences AI initiative on Hugging Face. The integration of these powerful tools enables our model to interpret and analyze complex medical data with unparalleled precision and depth. ### Dataset Overview MedMCQA Dataset: At the heart of our model's training lies the MedMCQA dataset, a comprehensive collection of over 194,000 multiple-choice questions (MCQs) derived from AIIMS & NEET PG medical entrance exams. This dataset spans an impressive array of 21 medical subjects, covering 2,400 healthcare topics, making it an invaluable resource for developing an AI adept at understanding and generating medical content. Each question in the MedMCQA dataset is meticulously designed to test a model's reasoning across various medical fields, providing not only the questions and correct answers but also detailed explanations and alternative options, enriching the model's learning process. ### Use Cases: Application in Medical Report Generation HL7 Medical Reports: Our AI model is fine-tuned to assist healthcare professionals by automating the creation of HL7 medical reports. HL7 (Health Level Seven) is a set of international standards for the transfer of clinical and administrative data between software applications used by various healthcare providers. By harnessing the MedMCQA dataset, our model is trained to understand the nuances and complexities of medical data, enabling it to generate detailed, accurate, and comprehensible medical reports. Diagnostic Support: Our AI model can analyze patient interactions and relevant health data to provide preliminary diagnostic insights, aiding doctors in their decision-making process. Medical Education: The model can serve as an advanced tool for medical students and professionals, offering detailed explanations and reasoning, akin to those found in the MedMCQA dataset, to enhance learning and understanding. Research and Analysis: By generating comprehensive reports, the model can assist medical researchers in compiling and analyzing data, facilitating advancements in medical research.
[ "MEDICAL DATA" ]
Manbehindthemadness/craft_mlt_25k
Manbehindthemadness
null
[ "license:mit", "region:us" ]
"2024-04-01T16:37:49Z"
2024-04-01T17:01:22+00:00
0
0
--- license: mit --- This is a re-publication of craft_mlt_25k from https://github.com/WindowsKonon1337/CRAFT-pytorch I have done this to preserve the model and provide an accessable location for future projects to acquire the file(s)
[ "CRAFT" ]
ostapeno/library-mistral7B_flan_10clsuters_3ep_lr1e-4
ostapeno
null
[ "region:us" ]
"2024-04-03T03:58:16Z"
2024-04-16T22:20:28+00:00
0
0
--- {} --- Number of experts present in the library: 10 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | c6o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text,cos_e_v1_11_question_description_option_text,cos_e_v1_11_question_option_description_id,qasc_qa_with_combined_facts_1,qasc_qa_with_separated_facts_1,qasc_qa_with_separated_facts_2,qasc_qa_with_separated_facts_3,qasc_qa_with_separated_facts_5,quarel_choose_between,quarel_heres_a_story,quarel_logic_test,quarel_testing_students,quartz_answer_question_based_on,quartz_answer_question_below,quartz_given_the_fact_answer_the_q,quartz_having_read_above_passage,quartz_paragraph_question_plain_concat,quartz_read_passage_below_choose,quartz_use_info_from_paragraph_question,quartz_use_info_from_question_paragraph,quoref_Answer_Question_Given_Context,ropes_background_new_situation_answer,ropes_background_situation_middle,ropes_given_background_situation,ropes_new_situation_background_answer,ropes_plain_background_situation,ropes_plain_bottom_hint,ropes_plain_no_background,ropes_prompt_beginning,ropes_prompt_bottom_hint_beginning,ropes_prompt_bottom_no_hint,ropes_prompt_mix,ropes_read_background_situation,sciq_Direct_Question_Closed_Book_,social_i_qa_Show_choices_and_generate_answer,wiqa_does_the_supposed_perturbation_have_an_effect,wiqa_effect_with_label_answer,wiqa_effect_with_string_answer,wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | c0o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question,adversarial_qa_dbidaf_generate_question,adversarial_qa_droberta_generate_question,app_reviews_generate_review,cot_creak,cot_esnli,cot_esnli_ii,dream_generate_first_utterance,dream_generate_last_utterance,duorc_ParaphraseRC_title_generation,duorc_SelfRC_title_generation,fix_punct,gem_common_gen_1_1_0,gem_dart_1_1_0,gigaword_1_2_0,huggingface_xsum,lambada_1_0_0,race_high_Write_a_multi_choice_question_for_the_following_article,race_high_Write_a_multi_choice_question_options_given_,race_middle_Write_a_multi_choice_question_for_the_following_article,race_middle_Write_a_multi_choice_question_options_given_,stream_aqua,stream_qed,wiqa_what_is_the_missing_first_step,wmt16_translate_fi_en_1_0_0,wmt16_translate_ro_en_1_0_0,yelp_polarity_reviews_0_2_0 | lora | | c8o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/anli_r1_0_1_0,anli_r2_0_1_0,anli_r3_0_1_0,cosmos_qa_1_0_0,cot_ecqa,cot_sensemaking,glue_cola_2_0_0,glue_mrpc_2_0_0,glue_sst2_2_0_0,imdb_reviews_plain_text_1_0_0,para_crawl_enes,super_glue_record_1_0_2,true_case,wmt14_translate_fr_en_1_0_0,wmt16_translate_de_en_1_0_0,word_segment | lora | | c9o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1,wiki_hop_original_choose_best_object_affirmative_2,wiki_hop_original_choose_best_object_affirmative_3,wiki_hop_original_choose_best_object_interrogative_1,wiki_hop_original_choose_best_object_interrogative_2,wiki_hop_original_explain_relation,wiki_hop_original_generate_object,wiki_hop_original_generate_subject,wiki_hop_original_generate_subject_and_object | lora | | c2o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review,cos_e_v1_11_question_option_description_text,cot_qasc,cot_strategyqa_ii,dbpedia_14_pick_one_category_for_the_following_text,definite_pronoun_resolution_1_1_0,kilt_tasks_hotpotqa_final_exam,math_dataset_algebra__linear_1d_1_0_0,qasc_qa_with_separated_facts_4,quarel_do_not_use,quoref_Context_Contains_Answer,race_high_Is_this_the_right_answer,race_middle_Is_this_the_right_answer,sciq_Direct_Question,sciq_Multiple_Choice,sciq_Multiple_Choice_Closed_Book_,sciq_Multiple_Choice_Question_First,social_i_qa_Show_choices_and_generate_index,stream_aqua_ii,super_glue_cb_1_0_2,super_glue_copa_1_0_2,unified_qa_science_inst,wiki_qa_Decide_good_answer,wiki_qa_Direct_Answer_to_Question,wiki_qa_Generate_Question_from_Topic,wiki_qa_Jeopardy_style,wiki_qa_Topic_Prediction_Answer_Only,wiki_qa_Topic_Prediction_Question_Only,wiki_qa_Topic_Prediction_Question_and_Answer_Pair,wiki_qa_automatic_system,wiki_qa_exercise,wiki_qa_found_on_google | lora | | c7o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/aeslc_1_0_0,cnn_dailymail_3_4_0,coqa_1_0_0,cot_gsm8k,dream_answer_to_dialogue,duorc_ParaphraseRC_build_story_around_qa,duorc_SelfRC_build_story_around_qa,gem_e2e_nlg_1_1_0,gem_web_nlg_en_1_1_0,gem_wiki_lingua_english_en_1_1_0,multi_news_1_0_0,wiki_bio_comprehension,wiki_bio_key_content,wiki_bio_what_content,wiki_bio_who,wiqa_what_is_the_final_step_of_the_following_process,wiqa_what_might_be_the_first_step_of_the_process,wiqa_what_might_be_the_last_step_of_the_process,wmt16_translate_tr_en_1_0_0 | lora | | c4o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question,duorc_ParaphraseRC_decide_worth_it,duorc_ParaphraseRC_extract_answer,duorc_ParaphraseRC_generate_question,duorc_ParaphraseRC_movie_director,duorc_ParaphraseRC_question_answering,duorc_SelfRC_answer_question,duorc_SelfRC_decide_worth_it,duorc_SelfRC_extract_answer,duorc_SelfRC_generate_question,duorc_SelfRC_movie_director,duorc_SelfRC_question_answering,quac_1_0_0,quoref_Answer_Friend_Question,quoref_Answer_Test,quoref_Find_Answer,quoref_Found_Context_Online,quoref_Given_Context_Answer_Question,quoref_Guess_Answer,quoref_Guess_Title_For_Context,quoref_Read_And_Extract_,quoref_What_Is_The_Answer | lora | | c3o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q,adversarial_qa_dbert_based_on,adversarial_qa_dbert_question_context_answer,adversarial_qa_dbert_tell_what_it_is,adversarial_qa_dbidaf_answer_the_following_q,adversarial_qa_dbidaf_based_on,adversarial_qa_dbidaf_question_context_answer,adversarial_qa_dbidaf_tell_what_it_is,adversarial_qa_droberta_answer_the_following_q,adversarial_qa_droberta_based_on,adversarial_qa_droberta_question_context_answer,adversarial_qa_droberta_tell_what_it_is,cos_e_v1_11_aligned_with_common_sense,cos_e_v1_11_explain_why_human,cos_e_v1_11_generate_explanation_given_text,cos_e_v1_11_i_think,cos_e_v1_11_rationale,drop_2_0_0,duorc_ParaphraseRC_generate_question_by_answer,duorc_SelfRC_generate_question_by_answer,kilt_tasks_hotpotqa_combining_facts,kilt_tasks_hotpotqa_formulate,kilt_tasks_hotpotqa_straighforward_qa,natural_questions_open_1_0_0,trivia_qa_rc_1_1_0,web_questions_get_the_answer,web_questions_potential_correct_answer,web_questions_question_answer,web_questions_short_general_knowledge_q,web_questions_whats_the_answer | lora | | c1o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/ag_news_subset_1_0_0,app_reviews_convert_to_rating,app_reviews_convert_to_star_rating,cot_creak_ii,cot_ecqa_ii,cot_gsm8k_ii,cot_sensemaking_ii,cot_strategyqa,dbpedia_14_given_a_choice_of_categories_,dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to,dbpedia_14_given_list_what_category_does_the_paragraph_belong_to,glue_mnli_2_0_0,glue_qnli_2_0_0,glue_qqp_2_0_0,glue_stsb_2_0_0,glue_wnli_2_0_0,kilt_tasks_hotpotqa_complex_question,paws_wiki_1_1_0,qasc_is_correct_1,qasc_is_correct_2,snli_1_1_0,social_i_qa_Check_if_a_random_answer_is_valid_or_not,social_i_qa_Generate_answer,social_i_qa_Generate_the_question_from_the_answer,social_i_qa_I_was_wondering,squad_v1_1_3_0_0,squad_v2_0_3_0_0,stream_qed_ii,super_glue_multirc_1_0_2,super_glue_rte_1_0_2,super_glue_wic_1_0_2,super_glue_wsc_fixed_1_0_2,trec_1_0_0,wiki_bio_guess_person,wiki_qa_Is_This_True_ | lora | | c5o10 | mistralai/Mistral-7B-v0.1 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id,cos_e_v1_11_question_description_option_id,dream_baseline,dream_read_the_following_conversation_and_answer_the_question,quail_context_description_question_answer_id,quail_context_description_question_answer_text,quail_context_description_question_text,quail_context_question_answer_description_id,quail_context_question_answer_description_text,quail_context_question_description_answer_id,quail_context_question_description_answer_text,quail_context_question_description_text,quail_description_context_question_answer_id,quail_description_context_question_answer_text,quail_description_context_question_text,quail_no_prompt_id,quail_no_prompt_text,race_high_Read_the_article_and_answer_the_question_no_option_,race_high_Select_the_best_answer,race_high_Select_the_best_answer_generate_span_,race_high_Select_the_best_answer_no_instructions_,race_high_Taking_a_test,race_middle_Read_the_article_and_answer_the_question_no_option_,race_middle_Select_the_best_answer,race_middle_Select_the_best_answer_generate_span_,race_middle_Select_the_best_answer_no_instructions_,race_middle_Taking_a_test | lora | Last updated on: 2024-04-06 04:06:45+00:00
[ "SCIQ" ]
steventango/GNorm2-docker
steventango
null
[ "region:us" ]
"2024-04-07T19:47:38Z"
2024-04-25T05:40:12+00:00
0
0
--- {} --- # GNorm2 *** GNorm2 is a gene name recognition and normalization tool with optimized functions and customizable configuration to the user preferences. The GNorm2 integrates multiple deep learning-based methods and achieves state-of-the-art performance. GNorm2 is freely available to download for stand-alone usage. [Download GNorm2 here](https://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/tmTools/download/GNorm2/GNorm2.tar.gz) ## Content - [Dependency package](#package) - [Introduction of folders](#intro) - [Running GNorm2](#pipeline) ## Dependency package <a name="package"></a> The codes have been tested using Python3.8/3.9 on CentOS and uses the following main dependencies on a CPU and GPU: - [TensorFlow 2.3.0](https://www.tensorflow.org/) - [Transformer 4.18.0](https://huggingface.co/docs/transformers/installation) - [stanza 1.4.0](stanfordnlp.github.io/stanza/) To install all dependencies automatically using the command: $ pip install -r requirements.txt ## Introduction of folders <a name="intro"></a> - src_python - GeneNER: the codes for gene recognition - SpeAss: the codes for species assignment - src_Java - GNormPluslib : the codes for gene normalization and species recogntion - GeneNER_SpeAss_run.py: the script for runing pipeline - GNormPlus.jar: the upgraded GNormPlus tools for gene normalization - gnorm_trained_models:pre-trianed models and trained NER/SA models - bioformer-cased-v1.0: the original bioformer model - BiomedNLP-PubMedBERT-base-uncased-abstract: the original pubmedbert model - geneNER - GeneNER-Bioformer/PubmedBERT-Allset.h5: the Gene NER models trained by all datasets - GeneNER-Bioformer/PubmedBERT-Trainset.h5: the Gene NER models trained by the training set only - SpeAss - SpeAss-Bioformer/PubmedBERT-SG-Allset.h5: the Species Assignment models trained by all datasets - SpeAss-Bioformer/PubmedBERT-SG-Trainset.h5: the Species Assignment models trained by the trianing set only - stanza - downloaded stanza library for offline usage - vocab: label files for the machine learning models of GeneNER and SpeAss - Dictionary: The dictionary folder contains all required files for gene normalization - CRF: CRF++ library (called by GNormPlus.sh) - Library: Ab3P library - tmp/tmp_GNR/tmp_SA/tmp_SR folders: temp folder - input/output folders: input and output folders. BioC (abstract or full text) and PubTator (abstract only) formats are both avaliable. - GNorm2.sh: the script to run GNorm2 - setup.GN.txt/setup.SR.txt/setup.txt the setup files for GNorm2. ## Running GNorm2 <a name="pipeline"></a> Please firstly download [GNorm2](https://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/tmTools/download/GNorm2/GNorm2.tar.gz) to your local. Below are the well-trained models (i.e., PubmedBERT/Bioformer) for Gene NER and Species Assignment. Models for Gene NER: - gnorm_trained_models/geneNER/GeneNER-PubmedBERT.h5 - gnorm_trained_models/geneNER/GeneNER-Bioformer.h5 Models for Species Assignment: - gnorm_trained_models/SpeAss/SpeAss-PubmedBERT.h5 - gnorm_trained_models/SpeAss/SpeAss-Bioformer.h5 The parameters of the input/output folders: - INPUT, default="input" - OUTPUT, default="output" [BioC-XML](bioc.sourceforge.net) or [PubTator](https://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/tmTools/Format.html) formats are both avaliabel to GNorm2. 1. Run GNorm2 Run Example: $ ./GNorm2.sh input output ## Acknowledgments This research was supported by the Intramural Research Program of the National Library of Medicine (NLM), National Institutes of Health.
[ "GNORMPLUS" ]
Zainn123/bioresults
Zainn123
null
[ "region:us" ]
"2024-04-10T09:58:32Z"
2024-04-10T09:58:45+00:00
0
0
--- {} --- hf (pretrained=johnsnowlabs/BioLing-7B-Dare), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 8 | Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| |-------------------------------|-------|------|-----:|--------|-----:|---|-----:| |stem |N/A |none | 0|acc |0.5164|± |0.0059| | | |none | 0|acc_norm|0.4705|± |0.0068| | - medmcqa |Yaml |none | 0|acc |0.4633|± |0.0077| | | |none | 0|acc_norm|0.4633|± |0.0077| | - medqa_4options |Yaml |none | 0|acc |0.4941|± |0.0140| | | |none | 0|acc_norm|0.4941|± |0.0140| | - anatomy (mmlu) | 0|none | 0|acc |0.5556|± |0.0429| | - clinical_knowledge (mmlu) | 0|none | 0|acc |0.6755|± |0.0288| | - college_biology (mmlu) | 0|none | 0|acc |0.6319|± |0.0403| | - college_medicine (mmlu) | 0|none | 0|acc |0.5896|± |0.0375| | - medical_genetics (mmlu) | 0|none | 0|acc |0.6900|± |0.0465| | - professional_medicine (mmlu)| 0|none | 0|acc |0.6654|± |0.0287| | - pubmedqa | 1|none | 0|acc |0.7480|± |0.0194| |Groups|Version|Filter|n-shot| Metric |Value | |Stderr| |------|-------|------|-----:|--------|-----:|---|-----:| |stem |N/A |none | 0|acc |0.5164|± |0.0059| | | |none | 0|acc_norm|0.4705|± |0.0068|
[ "MEDQA", "PUBMEDQA" ]
openkg/OneKE
openkg
null
[ "arxiv:2402.14710", "license:cc-by-nc-sa-4.0", "region:us" ]
"2024-04-11T04:43:48Z"
2024-04-11T06:05:36+00:00
0
17
--- license: cc-by-nc-sa-4.0 --- <p align="center"> <a href="https://github.com/zjunlp/deepke"> <img src="assets/oneke_logo.png" width="400"/></a> <p> <p align="center"> <a href="https://oneke.openkg.cn/"> <img alt="Documentation" src="https://img.shields.io/badge/demo-website-blue"> </a> <a href="https://pypi.org/project/deepke/#files"> <img alt="PyPI" src="https://img.shields.io/pypi/v/deepke"> </a> <a href="https://github.com/zjunlp/DeepKE/blob/master/LICENSE"> <img alt="GitHub" src="https://img.shields.io/github/license/zjunlp/deepke"> </a> <a href="http://zjunlp.github.io/DeepKE"> <img alt="Documentation" src="https://img.shields.io/badge/doc-website-red"> </a> </p> <h1 align="center"> <p>OneKE: A Bilingual Large Language Model for <br>Knowledge Extraction</p> </h1> - [What is OneKE?](#what-is-oneke) - [How is OneKE trained?](#how-is-oneke-trained) - [Getting Started with OneKE](#getting-started-with-oneke) - [Quick Start](#quick-start) - [Advanced Use of OneKE](#advanced-use-of-oneke) - [OneKE Instruction Format](#oneke-instruction-format) - [Conversion of OneKE Instruction Format](#conversion-of-oneke-instruction-format) - [Customized Schema Description Instructions](#customized-schema-description-instructions) - [Evaluation](#evaluation) - [Continue Training](#continue-training) - [Citation](#citation) ## What is OneKE? OneKE is a new bilingual knowledge extraction large model developed jointly by Zhejiang University and Ant Group, leveraging their years of accumulation in knowledge graph and natural language processing technology. Launched in 2024, the model employs schema-based polling instruction construction technology and is optimized to enhance the model's generalization capabilities for structured information extraction. <p align="center" width="100%"> <a href="" target="_blank"><img src="assets/oneke.gif" alt="OneKE" style="width: 100%; min-width: 20px; display: block; margin: auto;"></a> </p> ## How is OneKE trained? OneKE mainly focuses on schema-generalizable information extraction. Due to issues such as non-standard formats, noisy data, and lack of diversity in existing extraction instruction data, OneKE adopted techniques such as normalization and cleaning of extraction instructions, difficult negative sample collection, and schema-based batched instruction construction, as shown in the illustration. For more detailed information, refer to the paper "[IEPile: Unearthing Large-Scale Schema-Based Information Extraction Corpus](https://arxiv.org/abs/2402.14710) [[Github](https://github.com/zjunlp/IEPile)]". The zero-shot generalization comparison results of OneKE with other large models are as follows: * `NER-en`: CrossNER_AI, CrossNER_literature, CrossNER_music, CrossNER_politics, CrossNER_science * `NER-zh`: WEIBONER, boson * `RE-zh`: COAE2016, IPRE, SKE2020 * `RE-en`: FewRel, Wiki-ZSL * `EE-en`: CrudeOilNews, WikiEvents, RAMS * `EE-zh`: FewFC, CCF Law <p align="center" width="50%"> <a href="" target="_blank"><img src="assets/oneke_results.png" alt="OneKE" style="width: 50%; min-width: 20px; display: block; margin: auto;"></a> </p> Here's the translation of the provided text into English: ## Getting Started with OneKE ### Model Donwload [HuggingFace](https://huggingface.co/zjunlp/OneKE) ### Quick Start It is recommended to have at least **20GB of VRAM** for training and inferencing. ```python import torch from transformers import ( AutoConfig, AutoTokenizer, AutoModelForCausalLM, GenerationConfig, BitsAndBytesConfig ) model_path = 'zjunlp/OneKE' config = AutoConfig.from_pretrained(model_path, trust_remote_code=True) tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) # 4bit量化OneKE quantization_config=BitsAndBytesConfig( load_in_4bit=True, llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, bnb_4bit_compute_dtype=torch.bfloat16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type="nf4", ) model = AutoModelForCausalLM.from_pretrained( model_path, config=config, device_map="auto", quantization_config=quantization_config, torch_dtype=torch.bfloat16, trust_remote_code=True, ) model.eval() system_prompt = '<<SYS>>\nYou are a helpful assistant. 你是一个乐于助人的助手。\n<</SYS>>\n\n' sintruct = "{\"instruction\": \"You are an expert in named entity recognition. Please extract entities that match the schema definition from the input. Return an empty list if the entity type does not exist. Please respond in the format of a JSON string.\", \"schema\": [\"person\", \"organization\", \"else\", \"location\"], \"input\": \"284 Robert Allenby ( Australia ) 69 71 71 73 , Miguel Angel Martin ( Spain ) 75 70 71 68 ( Allenby won at first play-off hole )\"}" sintruct = '[INST] ' + system_prompt + sintruct + '[/INST]' input_ids = tokenizer.encode(sintruct, return_tensors="pt") input_length = input_ids.size(1) generation_output = model.generate(input_ids=input_ids, generation_config=GenerationConfig(max_length=1024, max_new_tokens=512, return_dict_in_generate=True)) generation_output = generation_output.sequences[0] generation_output = generation_output[input_length:] output = tokenizer.decode(generation_output, skip_special_tokens=True) print(output) ``` ### Advanced Use of OneKE ### OneKE Instruction Format The instructions in OneKE are formatted in a dictionary-type string similar to JSON. It consists of three fields: (1) **`'instruction'`**, which is the task description, specifies in natural language the role the model plays and the task to be completed; (2) **`'schema'`**, a list of labels to be extracted, clearly indicates the key fields of the information to be extracted, reflecting the user's needs, and is dynamic and changeable; (3) **`'input'`**, refers to the source text for information extraction. Below are examples of instructions for various tasks: <details> <summary><b>Named Entity Recognition (NER)</b></summary> ```json { "instruction": "You are an expert specializing in entity extraction. Please extract entities that comply with the schema definition from the input; return an empty list for non-existent entity types. Please respond in the JSON string format.", "schema": ["Person Name", "Education", "Position", "Nationality"], "input": "Mr. Liu Zhijian: Born in 1956, Chinese nationality, no permanent residency abroad, member of the Communist Party, associate degree, senior economist." } ``` </details> <details> <summary><b>Relation Extraction (RE)</b></summary> ```json { "instruction": "You are an expert specializing in relation extraction. Please extract relationship triples that comply with the schema definition from the input; return an empty list for non-existent relationships. Please respond in the JSON string format.", "schema": ["Father", "Husband", "Postal Code", "Mother"], "input": "Ding Long took out his life savings of $12,000, which without a doubt was a substantial amount at the end of the 19th century, plus Carpentier's donation, they both funded Columbia University's sinology research together." } ``` </details> <details> <summary><b>Knowledge Graph Construction (KGC)</b></summary> ```json { "instruction": "You are an expert in structuring knowledge about graph entities. Based on the schema description of the input entity type, extract the corresponding entity instances and their property information from the text; do not output non-existent properties, return a list if there are multiple values for a property, and provide the output in a parseable json format.", "schema": [ { "entity_type": "Person", "attributes": ["Chinese Name", "English Name", "Ancestral Home", "Date of Birth", "Place of Birth", "Occupation", "Alma Mater", "Works", "Awards"] } ], "input": "Jay Chou (Jay Chou), born on January 18, 1979, in New Taipei City, Taiwan Province, ancestral home in Yongchun County, Quanzhou City, Fujian Province, Chinese pop singer, musician, actor, director, screenwriter, graduated from Tamkang High School. In 2000, he released his debut album 'Jay'. In 2001, he cemented his style of blending Eastern and Western music with the album 'Fantasy'. In 2002, he held ‘The One’ world tour; the same year, he won the Best Composer award at the 13th Taiwan Golden Melody Awards with the song 'Love Before the Century'." } ``` </details> <details> <summary><b>Event Extraction (EE)</b></summary> ```json { "instruction": "You are an expert specializing in event extraction. Please extract events that match the defined schema from the input; return an empty list for non-existent events, NAN for non-existent arguments, and a list if there are multiple values for an argument. Please provide your response in JSON string format.", "schema": [ { "event_type": "Finance/Trading - Interest Rate Hike", "trigger": true, "arguments": [ "Time" ] }, { "event_type": "Finance/Trading - Interest Rate Cut", "trigger": true, "arguments": [ "Cut Magnitude" ] }, { "event_type": "Finance/Trading - Price Increase", "trigger": true, "arguments": [ "Price Raiser" ] }, { "event_type": "Finance/Trading - Price Cut", "trigger": true, "arguments": [ "Price Cutter", "Time" ] } ], "input": "AI risk control solution provider Vezetech secures tens of millions of dollars in Series C+ funding" } ``` </details> <details> <summary><b>Event Trigger Identification (EET)</b></summary> ```json { "instruction": "You are an expert specializing in event trigger identification. Please extract the event types and triggers that match the defined schema from the input; return an empty list if the event type doesn't exist. Please provide your response in JSON string format.", "schema": ["Organizational Relationship - Dissolve", "Organizational Relationship - Layoff", "Organizational Relationship - Dismiss", "Competition Behavior - Promotion"], "input": "Nestlé lays off 4,000 employees: When the times leave you behind, they won't even say goodbye!" } ``` </details> <details> <summary><b>Event Argument Extraction (EEA)</b></summary> ```json { "instruction": "You are an expert specializing in event argument extraction. Please extract the event arguments and their roles that match the defined schema from the input; return NAN or an empty dictionary for non-existent arguments, and a list if there are multiple values for an argument. Please provide your response in JSON string format.", "schema": [{"event_type": "Organizational Relationship - Resignation/Departure", "arguments": ["Resigner", "Time", "Former Organization"]}], "input": "Nestlé lays off 4,000 employees: When the times leave you behind, they won't even say goodbye!" } ``` </details> ### Conversion of OneKE Instruction Format **List of Instructions**: ```python instruction_mapper = { 'NERzh': "你是专门进行实体抽取的专家。请从input中抽取出符合schema定义的实体,不存在的实体类型返回空列表。请按照JSON字符串的格式回答。", 'REzh': "你是专门进行关系抽取的专家。请从input中抽取出符合schema定义的关系三元组,不存在的关系返回空列表。请按照JSON字符串的格式回答。", 'EEzh': "你是专门进行事件提取的专家。请从input中抽取出符合schema定义的事件,不存在的事件返回空列表,不存在的论元返回NAN,如果论元存在多值请返回列表。请按照JSON字符串的格式回答。", 'EETzh': "你是专门进行事件提取的专家。请从input中抽取出符合schema定义的事件类型及事件触发词,不存在的事件返回空列表。请按照JSON字符串的格式回答。", 'EEAzh': "你是专门进行事件论元提取的专家。请从input中抽取出符合schema定义的事件论元及论元角色,不存在的论元返回NAN或空字典,如果论元存在多值请返回列表。请按照JSON字符串的格式回答。", 'KGzh': '你是一个图谱实体知识结构化专家。根据输入实体类型(entity type)的schema描述,从文本中抽取出相应的实体实例和其属性信息,不存在的属性不输出, 属性存在多值就返回列表,并输出为可解析的json格式。', 'NERen': "You are an expert in named entity recognition. Please extract entities that match the schema definition from the input. Return an empty list if the entity type does not exist. Please respond in the format of a JSON string.", 'REen': "You are an expert in relationship extraction. Please extract relationship triples that match the schema definition from the input. Return an empty list for relationships that do not exist. Please respond in the format of a JSON string.", 'EEen': "You are an expert in event extraction. Please extract events from the input that conform to the schema definition. Return an empty list for events that do not exist, and return NAN for arguments that do not exist. If an argument has multiple values, please return a list. Respond in the format of a JSON string.", 'EETen': "You are an expert in event extraction. Please extract event types and event trigger words from the input that conform to the schema definition. Return an empty list for non-existent events. Please respond in the format of a JSON string.", 'EEAen': "You are an expert in event argument extraction. Please extract event arguments and their roles from the input that conform to the schema definition, which already includes event trigger words. If an argument does not exist, return NAN or an empty dictionary. Please respond in the format of a JSON string.", 'KGen': 'You are an expert in structured knowledge systems for graph entities. Based on the schema description of the input entity type, you extract the corresponding entity instances and their attribute information from the text. Attributes that do not exist should not be output. If an attribute has multiple values, a list should be returned. The results should be output in a parsable JSON format.', } ``` Recommended **Split Numbers** for Each Task: ```python split_num_mapper = { 'NER':6, 'RE':4, 'EE':4, 'EET':4, 'EEA':4, 'KG':1 } ``` Since predicting all schemas in the label set at once is too challenging and not easily scalable, OneKE uses a batched approach during training. It divides the number of schemas asked in the instructions, querying a fixed number of schemas at a time. Hence, if the label set of a piece of data is too long, it will be split into multiple instructions that the model will address in turns. **schema格式**: ```python NER: ["Person Name", "Education", "Position", "Nationality"] # List of strings RE: ["Father", "Husband", "Postal Code", "Mother"] # List of strings EE: [{"event_type": "Finance/Trading - Interest Rate Hike", "trigger": True, "arguments": ["Time"]}, {"event_type": "Finance/Trading - Interest Rate Cut", "trigger": True, "arguments": ["Cut Magnitude"]}] # List of dictionaries, "event_type" is a string, "trigger" is a bool, "arguments" is a list EET: ["Organizational Relationship - Dissolution", "Organizational Relationship - Layoff", "Organizational Relationship - Dismissal", "Competition Behavior - Advancement"] # List of strings EEA: [{"event_type": "Finance/Trading - Interest Rate Hike", "arguments": ["Time"]}, {"event_type": "Finance/Trading - Interest Rate Cut", "arguments": ["Cut Magnitude"]}] # List of dictionaries, "event_type" is a string, "arguments" is a list ``` Below is a simple Batched Instruction Generation script: ```python def get_instruction(language, task, schema, input): sintructs = [] split_num = split_num_mapper[task] if type(schema) == dict: sintruct = json.dumps({'instruction':instruction_mapper[task+language], 'schema':schema, 'input':input}, ensure_ascii=False) sintructs.append(sintruct) else: split_schemas = [schema[i:i+split_num] for i in range(0, len(schema), split_num)] for split_schema in split_schemas: sintruct = json.dumps({'instruction':instruction_mapper[task+language], 'schema':split_schema, 'input':input}, ensure_ascii=False) sintructs.append(sintruct) return sintructs ``` For more detailed data conversion, please refer to [InstructKGC/README_CN.md/2.3 Testing Data Conversion](./InstructKGC/README_CN.md/#23测试数据转换). Below is an example using the aforementioned simple script: ```python task = 'NER' language = 'en' schema = ['person', 'organization', 'else', 'location'] split_num = split_num_mapper[task] split_schemas = [schema[i:i+split_num] for i in range(0, len(schema), split_num)] input = '284 Robert Allenby ( Australia ) 69 71 71 73 , Miguel Angel Martin ( Spain ) 75 70 71 68 ( Allenby won at first play-off hole )' sintructs = [] for split_schema in split_schemas: sintruct = json.dumps({'instruction':instruction_mapper[task+language], 'schema':split_schema, 'input':input}, ensure_ascii=False) sintructs.append(sintruct) ``` > '{"instruction": "You are an expert in named entity recognition. Please extract entities that match the schema definition from the input. Return an empty list if the entity type does not exist. Please respond in the format of a JSON string.", "schema": ["person", "organization", "else", "location"], "input": "284 Robert Allenby ( Australia ) 69 71 71 73 , Miguel Angel Martin ( Spain ) 75 70 71 68 ( Allenby won at first play-off hole )"}' ### Customized Schema Description Instructions ```json { "instruction": "You are an expert specializing in entity extraction. Please extract entities that comply with the defined schema from the input; return an empty list for non-existent entity types. Please respond in JSON string format.", "schema": { "Position": "The entity type describes the occupation or official position of an individual or group, including specific role names such as 'producer', 'scorekeeper', 'ascetic', 'oil painter'.", "Attraction": "The entity type of attraction includes buildings, museums, memorials, art galleries, rivers, peaks, etc. Representative entities include the Pentagon, Tate Modern, Zheng Chenggong Memorial Hall, Duxi Palace, Barikasa, Robo River, Gunung Batur, Yugong Yishan LIVE, Xu Beihong Memorial Hall, Madame Tussauds, etc.", "Company": "Company is an entity type representing any legal entity or business organization. This type of entity can be a catering group, manufacturer, retailer, hotel, bank, design institute, etc. Examples include: 'Shangri-La Hotel Group', 'JVC', 'Shanghai Coolray Professional eSports Peripheral Store', 'K2&bull;Haitang Bay', 'Wuhan Iron and Steel', 'louisvuitton', 'Bank of Scotland', 'Beijing Institute of Architectural Design', '7 Days Inn', 'Vanke Group'.", "Address": "Address entities refer to entities with geographical location information, representing specific places such as a country, city, region, street, or abstract geographic areas. Examples include: 'the river dock at the southeast tip of downtown Manhattan', 'Tuapse', 'Venice, Italy', 'Huzhou Hot Spring Golf Course', 'North Carolina', 'Beijing-Tianjin region', 'Happy Internet Cafe', 'Yinian Nursing Home', 'Shangtang Town Pudong', 'Inner Mongolia Autonomous Region Chifeng City', etc.", "Organization": "Organizational entities refer to collective organizations such as companies, shops, clubs, schools, etc. They play a certain role in social and economic activities and have certain personality rights.", "Movie": "Movie entities include titles of movies in Chinese or English, and sometimes also include names of characters in films." }, "input": "It is difficult for me to imagine setting up another Haifishing Plaza. When we obtained this project, I just happened to be in Sanya." } ``` <details> <summary><b>Relation Extraction (RE) Description Instructions</b></summary> ```json { "instruction": "You are an expert specializing in relation extraction. Please extract triples that match the defined schema from the input; return an empty list for non-existent relations. Please respond in JSON string format.", "schema": { "Ethnicity": "Ethnicity", "Alma Mater": "This type of relationship describes the connection between a person and their alma mater; the person is the subject, and the alma mater is the object. By identifying the names of people and schools in the text and analyzing the relationship of graduation between them based on word combinations and contextual information.", "Lead Actor": "This is a type of relationship that describes the connection between a film or television work and its main actors; the subject is the film or television work and the object is the actor. In a valid 'Lead Actor' relationship, the actor (object) plays an important role in the work (subject).", "Father": "This type of relationship is used to indicate the kinship between a father and a child, where the father is the birth parent or caregiver of the child. In the triple, the subject of the 'Father' relation type is the child, and the object is the father." }, "input": "Throughout history, all those who have portrayed the character 'Chu Liuxiang' from Gu Long's novels are recognized as handsome men in the entertainment industry. In 2011, 36-year-old Zhang Zhiyao played Chu Liuxiang in 'The New Adventures of Chu Liuxiang', remaining irresistibly handsome." } ``` </details> <details> <summary><b>Event Extraction (EE) Description Instructions</b></summary> ```json { "instruction": "You are an expert specializing in event extraction. Please extract events that match the schema definition from the input; return an empty list for non-existent events, NAN for non-existent arguments, and a list if there are multiple values for an argument. Please respond in JSON string format.", "schema": { "Finance/Trading - Listing": { "Finance/Trading - Listing": "The act of a financial entity being listed on the stock market mainly involves companies, stocks, etc. Positive examples include specific information about a company or stock listing, while negative examples are unrelated to such activities.", "trigger": true, "arguments": { "Financing Amount": "Refers to the total amount of funds raised by a company in a listing event. It sums up the revenue of all share issues and is measured in currency, including but not limited to units like 'billion', 'million', 'dollars', 'RMB', etc.", "Time": "Describes the specific time of the listing event, which can be a specific date or relative time, and may also include location information and specific days and weeks.", "Listing Enterprise": "Refers to the company or enterprise that is conducting an IPO or has already been listed on the trading market in a listing event. Examples include: 'Shanghai Henlius Biotech', 'Three Squirrels', 'Baoxin Software', 'Little Bear Electric', 'Jinshang Bank', 'Beyond Meat (BYND)', 'DouYu gaming live-streaming platform', 'fast food empire', and 'autonomous driving lidar manufacturer Velodyne', etc.", "Location": "The specific location of the financial or trading event, such as a city, building, or room." } }, "Organizational Relationship - Resignation/Departure": { "Organizational Relationship - Resignation/Departure": "The event type 'Organizational Relationship - Resignation/Departure' refers to changes in the relationship between individuals or organizational members and their organization, mainly including 'resignation', 'requesting to resign', 'stepping down', 'leaving the team', 'retirement', 'leaving', etc. Often occurs in scenarios of high-level personnel changes, government officials changes, or athletes transfers. Examples: 'Li Nan announced resignation', 'Yu Xubo resigned from the position of chairman of the board just three months after taking office, Chen Lang succeeded'.", "trigger": true, "arguments": { "Resigner": "Refers to the individual or group who actively or passively leaves their original position or job post in an organizational relationship resignation/departure event. It can be one person or a group of people, such as: 'Finance Minister', '90s born guy from Shaoyang Longhui, Ouyang En and', 'Xiong Xiaoge', '*ST Changsheng two deputy general managers', 'Yang Tao', 'pilot Ma Qiang', 'HE WEI', '5 Baidu executives', 'Youxin Group COO Peng Weilian', 'Jianke Institute securities representative Shu Yanming', etc.", "Time": "Indicates the specific point in time or period when the resignation/departure event occurred, generally including specific dates, weeks, times, etc., like 'September 19', 'the evening of June 29', 'this Saturday', '10:30 AM on July 9', 'the morning of June 12', 'April 9', 'September 10', 'local time on Sunday', 'September 12', '10 AM on October 15', etc." } }, "Finance/Trading - Interest Rate Increase": { "Finance/Trading - Interest Rate Increase": "This event describes banks or financial institutions raising interest rates to tighten the money supply. The typical trigger word is 'hike'. 'Hike' indicates the occurrence of the Finance/Trading - Interest Rate Increase event.", "trigger": true, "arguments": { "Rate of Increase": "The rate of increase is usually presented as a percentage or basis points, indicating the degree or range of the interest rate hike in the event. Examples include: 'to 5.75%', '25 basis points', 'the benchmark rate from 0.25% up to 0.5%', '25 basis points'.", "Hiking Institution": "The hiking institution is the financial institution with the authority to determine or implement the interest rate hike policy in a Finance/Trading - Interest Rate Increase event, such as central banks from different countries (e.g., Bank of England, Federal Reserve, European Central Bank) or financial institutions (e.g., Bank of England).", "Time": "Indicates the specific date or time period when the Finance/Trading - Interest Rate Increase event occurred, such as 'the morning of June 18th', 'January 24th', 'three months later', etc. The specific expression includes time accurate to the minute, such as '11:00 on December 28, 2018', relative time, such as 'yesterday (2nd)', and special time expressions like 'Mid-Autumn Festival'." } }, "Organizational Relationship - Contract Termination": { "Organizational Relationship - Contract Termination": "Situations of contract cancellation or termination usually occur in the business, entertainment, or sports domains. Trigger words include 'leave', 'trade', 'cut', 'contract expiry', 'contract termination', 'sell-off', 'release', 'send out', 'contract break', etc. Positive examples include 'Peng Yuchang terminates his contract' and 'Jiang Mengjie nearly bankrupt after contract termination'. Negative examples are like 'Federer withdrew from the competition'.", "trigger": true, "arguments": { "Party Being Terminated": "In an organizational relationship contract termination event, the role is the party whose agreement or contract relation is being dissolved, and might be an individual or an organization, such as an athlete, film producer, company, etc. For instance, 'seven-time All-Star Joe Johnson', 'the production side of 'A Little Wish'', 'Raptors', 'Samsung', etc." } } }, "input": "News from August 20th, according to Tencent News 'Frontline' report, informed sources stated that in order to control cost expenditure, NIO plans to reduce the number of staff at its U.S. branch, excluding those involved in the autonomous driving business, to about 200. As of August 16th, U.S. time, NIO's Silicon Valley branch had cut 100 employees." } ``` </details> <details> <summary><b>Knowledge Graph Construction (KGC) Description Instructions</b></summary> ```json { "instruction": "You are an expert in structuring knowledge about graph entities. Based on the schema description for the input entity type, extract the corresponding entity instances and their attribute information from the text; do not output non-existent attributes, return a list for attributes with multiple values, and provide the output in a parseable JSON format.", "schema": [ { "entity_type": "Person", "attributes": { "Chinese Name": "The Chinese name of the person", "English Name": "The English name of the person", "Ancestral Home": "The ancestral address of the person", "Date of Birth": "Birthday, birth date", "Place of Birth": "The place of birth, administrative region", "Occupation": "The occupation, position, identity of the person", "Alma Mater": "The middle school, university, college from which the person graduated", "Works": "Albums, songs, novels, published books, participated film and television works, etc.", "Awards": "Various awards and honors received by the person" } } ], "input": "Jay Chou (Jay Chou), born on January 18, 1979, in New Taipei City, Taiwan Province, with ancestral home in Yongchun County, Quanzhou City, Fujian Province, is a Chinese pop musician, actor, director, and screenwriter. He graduated from Tamkang High School. In 2000, he released his debut music album 'Jay.' In 2001, he cemented his fusion style of Eastern and Western music with the album 'Fantasy.' In 2002, he held 'The One' world tour; that same year, he won the Best Composer award at the 13th Taiwan Golden Melody Awards for the song 'Love Before the Century.'" } ``` </details> ## Evaluation To extract structured content from the output text and to assess it, please refer to [InstructKGC/README_CN.md/7. Evaluation](./InstructKGC/README_CN.md/#🧾-7评估). ## Continue Training To continue training OneKE, refer to [InstructKGC/4.9 Domain-specific Data Continual Training](./InstructKGC/README_CN.md/#49领域内数据继续训练). ## Citation If you have used OneKE in your work, please kindly cite the following paper: ```bibtex @article{DBLP:journals/corr/abs-2402-14710, author = {Honghao Gui and Lin Yuan and Hongbin Ye and Ningyu Zhang and Mengshu Sun and Lei Liang and Huajun Chen}, title = {IEPile: Unearthing Large-Scale Schema-Based Information Extraction Corpus}, journal = {CoRR}, volume = {abs/2402.14710}, year = {2024}, url = {https://doi.org/10.48550/arXiv.2402.14710}, doi = {10.48550/ARXIV.2402.14710}, eprinttype = {arXiv}, eprint = {2402.14710}, timestamp = {Tue, 09 Apr 2024 07:32:43 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2402-14710.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
[ "BEAR" ]
aistrosight/Neg-CamemBERT-bio
aistrosight
token-classification
[ "transformers", "pytorch", "camembert", "token-classification", "medical", "fr", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
"2024-04-12T12:33:19Z"
2024-04-30T11:43:44+00:00
0
0
--- language: - fr library_name: transformers license: mit metrics: - precision - recall - f1 pipeline_tag: token-classification tags: - medical widget: - text: absence de signe d'anomalie des petites voies aériennes, notamment pas de signe de piégeage. --- # Neg-CamemBERT-bio: Language model for negation detection in radiological and other clinical texts for the French language. Neg-CamemBERT-bio is a refined version of the transformer-based [**CamemBERT-bio-base model**](https://huggingface.co/almanach/camembert-bio-base), fine-tuned for the recognition of negations in clinical texts. Neg-CamemBERT-bio automatically detects both the negation cues and their scope. A public model is available for negation recognition in biomedical texts, along with two additional models that are kept private since trained with potentially sensitive data from Lyon University Hospital (Hospices Civils de Lyoon, HCL). — one dedicated to radiology reports and the other designed more broadly for various medical texts. ## Model Details 1- Neg-CamemBERT-bio: Fine-tuning of the CamemBERT-bio-base model for negation recognition in biomedical texts in French. 2- Neg-Radio-CamemBERT-bio: Fine-tuning of the CamemBERT-bio-base model for negation recognition in anonymized texts extracted from French-written thoracic CT scans provided by the Radiology Department of the Hospices Civils of Lyon. 3- Neg-Medical-CamemBERT-bio: Fine-tuning of the CamemBERT-bio-base model for negation recognition in clinical texts across various medical domains (Radiology, Biomedical, Medical literature,...) in French. | Model name | Type | Corpus train | Number of sentences | Negative sentences | | :----------- | :---- | :---- | :---------------: | :-------------------: | | `Neg-CamemBERT-bio` | public| ESSAI + CAS | 11 037 | 1 812 |2 009 | | `Neg-Radio-CamemBERT-bio`| privite| RADIO | 10 798 | 2 321 | 2 762| | `Neg-Medical-CamemBERT-bio` | privite | RADIO + ESSAI + CAS + QUAERO| 21 956 | 4 244| ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> | **Corpus** | **Details** |**Licence**| |------------|--------------------------------------------------------------------|--------------| | RADIO | Corpus extracted from thoracic scan reports provided by the Radiology Department of Lyon University Hospital (Hospices Civils de Lyon, HCL).| Private corpus| | ESSAI | ESSAI Clinical is a freely-available corpus that contains clinical trial protocols in French language collected from the registry of the National Cancer Institute (data available [here](https://clementdalloux.fr/?page_id=28)). | CC BY-NC-SA 4.0 DEED | | CAS | CAS is a freely-available corpus in French containing clinical cases as published in scientific, legal, or educational literature (data available [here](https://clementdalloux.fr/?page_id=28)).| CC BY-NC-SA 4.0 DEED | | QUAERO| QUAERO is freely-available corpus that contains a vast amount of information in the biomedical field and is available in the form of free-text in natural language (data available [here](https://quaerofrenchmed.limsi.fr/)). | GNU Free Documentation License | | 413 M | The training dataset distinguishes between the beginning, inside, and end of each negation entity using a BIO annotation scheme. Abbreviation|Description -|- O |Outside of a named entity, represents the affirmative part of the sentence. B-cue |Beginning of the negation cue. I-cue | Inside of the negation cue. B-scope |Beginning of the negation scope. I-scope |Inside of the negation scope. ### Fine-tuning The CamemBERT-bio-base model, along with its tokenizer, was fine-tuned for the token classification task in order to identify negation cues and their scope. We used the Adam optimizer with a learning rate of 5e−5, and a batch size of 16. The maximum length limit was set to 512 tokens. The model was trained using cross-validation with 10 folds (90% training / 10% validation). For each fold, the model was trained for 25 epochs. We eventually selected the best model out of the 10 folds. ### Evaluation To evaluate the performance of the model and quantify the results, we used the following metrics: precision (P), recall (R), and the F1 score (F1). The scores were measured using the seqeval tool. ### Results Validation was performed on a 10% sample of sentences from the training set for each model. | Model's | Validation Dataset | Metrcis Score | Entity 1: cue | Entity 2: scope | | :----------- | :------:| :---------------:|:---------------: | :---------------: | | `Neg-CamemBERT-bio`| 10% (ESSAI+CAS) | P | 95.70 ± 0.94| 86.43 ± 1.12| | | | R | 97.70 ± 0.56| 85.55 ± 1.33| | | | F1 | 96.68 ± 0.46 |87.46 ± 0.93 | | `Neg-Radio-CamemBERT-bio` | 10% (RADIO) | P | 99.35 ± 0.24 | 94.19 ± 0.94| | | | R |99.37 ± 0.31 |94.80 ± 0.84 | | | | F1 |99.36 ± 0.25 |94.49 ± 0.80 | | `Neg-Medical-CamemBERT-bio`| 10%(NegRADIO + ESSAI + CAS + QUAERO)| P | 97.75 ± 0.45 |90.48 ± 0.74 | | | | R |98.67 ± 0.20 |91.34 ± 0.60 | | | | F1 |98.20 ± 0.21 | 90.90 ± 0.585| ### Model Description - **Developed by:** Salim SADOUNE, Antoine Richard, François Talbot,Thomas Guyet, Loic Boussel and Hugues Berry - **Model type:** NER - **Language(s) (NLP):** French - **License:** MIT - **Finetuned from model:** See the [**CamemBERT-bio-base model**](https://huggingface.co/almanach/camembert-bio-base) for more information on this model. ### Direct Use You can use the public model with Transformers *pipeline* for NER. ```python from transformers import AutoTokenizer, AutoModelForTokenClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained("aistrosight/Neg-CamemBERT-bio") model = AutoModelForTokenClassification.from_pretrained("aistrosight/Neg-CamemBERT-bio") NegNamedEntityRecogniser = pipeline(task="token-classification", model=model , tokenizer = tokenizer, aggregation_strategy="simple") text = ["absence de signe d'anomalie des petites voies aériennes, notamment pas de signe de piégeage."] sample =NegNamedEntityRecogniser(text) print(sample) ``` You can visualize the cue and the scope of the negation with the library spacy in a Jupyter notebook. ```python import spacy def visualize(sample, text): colors = {'scope_neg': "#61ffab", "cue_neg": "#ff6961"} options = {"ents": ['scope_neg', 'cue_neg'], "colors": colors} for i in range(len(sample)): entities = [] for ents in sample[i]: entities.append({"end": ents["end"], "label": ents["entity_group"], "start": ents["start"]}) displacy.render({"ents": entities,"text": text[i]}, style="ent", manual=True,options=options, jupyter=True) visualize(sample,text) ``` #### Limitations and bias The capacity of the Neg-CamemBERT-bio model is constrained due to the limited size of its training set, which includes a restricted number of examples for certain negation indicators *("sauf", "jamais", "hormis",...)* that appear less frequently. This limitation poses challenges for generalizing to other cases. The tokenizer, not specifically designed for radiology, can lead to confusion for the Neg-Radio-CamemBERT-bio model when predicting the scope. It is important to note that both the Radio corpus and the ESSAI + CAS corpus have been annotated by different annotators on purpose, which may introduce confusion into the ability of the Neg-Medical-CamemBERT-bio model to predict this scope accurately.
[ "CAS", "ESSAI", "QUAERO" ]
zhan1993/private_library_phi2_epoch_0
zhan1993
null
[ "region:us" ]
"2024-04-15T12:10:10Z"
2024-04-19T14:39:15+00:00
0
0
--- {} --- Number of experts present in the library: 263 | Expert Name | Base Model | Trained on | Adapter Type | | --- | --- | --- | --- | | sciq_Multiple_Choice | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice | lora | | wiki_hop_original_choose_best_object_interrogative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_1 | lora | | squad_v2_0_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v2_0_3_0_0 | lora | | wiki_qa_exercise | phi-2 | sordonia/flan-10k-flat/wiki_qa_exercise | lora | | race_high_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_high_Taking_a_test | lora | | adversarial_qa_dbert_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_generate_question | lora | | quoref_Found_Context_Online | phi-2 | sordonia/flan-10k-flat/quoref_Found_Context_Online | lora | | web_questions_get_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_get_the_answer | lora | | duorc_SelfRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question_by_answer | lora | | quarel_testing_students | phi-2 | sordonia/flan-10k-flat/quarel_testing_students | lora | | qasc_qa_with_separated_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_1 | lora | | wiki_qa_Is_This_True_ | phi-2 | sordonia/flan-10k-flat/wiki_qa_Is_This_True_ | lora | | race_high_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_high_Read_the_article_and_answer_the_question_no_option_ | lora | | cot_gsm8k_ii | phi-2 | sordonia/flan-10k-flat/cot_gsm8k_ii | lora | | gem_wiki_lingua_english_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_wiki_lingua_english_en_1_1_0 | lora | | unified_qa_science_inst | phi-2 | sordonia/flan-10k-flat/unified_qa_science_inst | lora | | quartz_use_info_from_paragraph_question | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_paragraph_question | lora | | wiki_hop_original_generate_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_object | lora | | quoref_What_Is_The_Answer | phi-2 | sordonia/flan-10k-flat/quoref_What_Is_The_Answer | lora | | adversarial_qa_droberta_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_generate_question | lora | | wiki_bio_comprehension | phi-2 | sordonia/flan-10k-flat/wiki_bio_comprehension | lora | | adversarial_qa_dbidaf_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_question_context_answer | lora | | wiki_bio_what_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_what_content | lora | | web_questions_whats_the_answer | phi-2 | sordonia/flan-10k-flat/web_questions_whats_the_answer | lora | | wiqa_what_is_the_missing_first_step | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_missing_first_step | lora | | adversarial_qa_droberta_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_question_context_answer | lora | | ropes_plain_bottom_hint | phi-2 | sordonia/flan-10k-flat/ropes_plain_bottom_hint | lora | | kilt_tasks_hotpotqa_combining_facts | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_combining_facts | lora | | cos_e_v1_11_aligned_with_common_sense | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_aligned_with_common_sense | lora | | gem_web_nlg_en_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_web_nlg_en_1_1_0 | lora | | web_questions_potential_correct_answer | phi-2 | sordonia/flan-10k-flat/web_questions_potential_correct_answer | lora | | wiki_qa_found_on_google | phi-2 | sordonia/flan-10k-flat/wiki_qa_found_on_google | lora | | duorc_ParaphraseRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_extract_answer | lora | | wmt16_translate_de_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_de_en_1_0_0 | lora | | quail_no_prompt_id | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_id | lora | | quoref_Guess_Title_For_Context | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Title_For_Context | lora | | duorc_SelfRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_decide_worth_it | lora | | ropes_prompt_mix | phi-2 | sordonia/flan-10k-flat/ropes_prompt_mix | lora | | adversarial_qa_droberta_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_tell_what_it_is | lora | | quail_context_question_answer_description_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_id | lora | | gem_common_gen_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_common_gen_1_1_0 | lora | | duorc_ParaphraseRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_answer_question | lora | | super_glue_cb_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_cb_1_0_2 | lora | | cnn_dailymail_3_4_0 | phi-2 | sordonia/flan-10k-flat/cnn_dailymail_3_4_0 | lora | | race_high_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_options_given_ | lora | | winogrande_1_1_0 | phi-2 | sordonia/flan-10k-flat/winogrande_1_1_0 | lora | | duorc_SelfRC_extract_answer | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_extract_answer | lora | | trec_1_0_0 | phi-2 | sordonia/flan-10k-flat/trec_1_0_0 | lora | | yelp_polarity_reviews_0_2_0 | phi-2 | sordonia/flan-10k-flat/yelp_polarity_reviews_0_2_0 | lora | | race_high_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer | lora | | para_crawl_enes | phi-2 | sordonia/flan-10k-flat/para_crawl_enes | lora | | qasc_is_correct_1 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_1 | lora | | app_reviews_generate_review | phi-2 | sordonia/flan-10k-flat/app_reviews_generate_review | lora | | ropes_read_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_read_background_situation | lora | | dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to | lora | | stream_aqua | phi-2 | sordonia/flan-10k-flat/stream_aqua | lora | | drop_2_0_0 | phi-2 | sordonia/flan-10k-flat/drop_2_0_0 | lora | | wiki_hop_original_choose_best_object_affirmative_1 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_1 | lora | | adversarial_qa_dbidaf_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_answer_the_following_q | lora | | social_i_qa_Generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_answer | lora | | stream_aqua_ii | phi-2 | sordonia/flan-10k-flat/stream_aqua_ii | lora | | glue_sst2_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_sst2_2_0_0 | lora | | cot_esnli | phi-2 | sordonia/flan-10k-flat/cot_esnli | lora | | race_high_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_no_instructions_ | lora | | duorc_SelfRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_build_story_around_qa | lora | | cot_esnli_ii | phi-2 | sordonia/flan-10k-flat/cot_esnli_ii | lora | | quail_no_prompt_text | phi-2 | sordonia/flan-10k-flat/quail_no_prompt_text | lora | | ropes_given_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_given_background_situation | lora | | quarel_logic_test | phi-2 | sordonia/flan-10k-flat/quarel_logic_test | lora | | adversarial_qa_dbidaf_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_based_on | lora | | super_glue_copa_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_copa_1_0_2 | lora | | cos_e_v1_11_i_think | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_i_think | lora | | quail_context_question_description_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_text | lora | | math_dataset_algebra__linear_1d_1_0_0 | phi-2 | sordonia/flan-10k-flat/math_dataset_algebra__linear_1d_1_0_0 | lora | | cosmos_qa_1_0_0 | phi-2 | sordonia/flan-10k-flat/cosmos_qa_1_0_0 | lora | | wiqa_effect_with_label_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_label_answer | lora | | app_reviews_convert_to_star_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_star_rating | lora | | qasc_qa_with_separated_facts_2 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_2 | lora | | race_middle_Select_the_best_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer | lora | | quartz_having_read_above_passage | phi-2 | sordonia/flan-10k-flat/quartz_having_read_above_passage | lora | | glue_qqp_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qqp_2_0_0 | lora | | cos_e_v1_11_question_description_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_id | lora | | cos_e_v1_11_question_option_description_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_text | lora | | imdb_reviews_plain_text_1_0_0 | phi-2 | sordonia/flan-10k-flat/imdb_reviews_plain_text_1_0_0 | lora | | wiki_hop_original_choose_best_object_affirmative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_2 | lora | | natural_questions_open_1_0_0 | phi-2 | sordonia/flan-10k-flat/natural_questions_open_1_0_0 | lora | | wiqa_effect_with_string_answer | phi-2 | sordonia/flan-10k-flat/wiqa_effect_with_string_answer | lora | | cos_e_v1_11_rationale | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_rationale | lora | | race_middle_Write_a_multi_choice_question_options_given_ | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_options_given_ | lora | | wiki_bio_guess_person | phi-2 | sordonia/flan-10k-flat/wiki_bio_guess_person | lora | | hellaswag_1_1_0 | phi-2 | sordonia/flan-10k-flat/hellaswag_1_1_0 | lora | | wiqa_does_the_supposed_perturbation_have_an_effect | phi-2 | sordonia/flan-10k-flat/wiqa_does_the_supposed_perturbation_have_an_effect | lora | | trivia_qa_rc_1_1_0 | phi-2 | sordonia/flan-10k-flat/trivia_qa_rc_1_1_0 | lora | | lambada_1_0_0 | phi-2 | sordonia/flan-10k-flat/lambada_1_0_0 | lora | | quoref_Read_And_Extract_ | phi-2 | sordonia/flan-10k-flat/quoref_Read_And_Extract_ | lora | | quail_context_description_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_id | lora | | quail_context_description_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_answer_text | lora | | duorc_SelfRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_question_answering | lora | | cot_sensemaking_ii | phi-2 | sordonia/flan-10k-flat/cot_sensemaking_ii | lora | | fix_punct | phi-2 | sordonia/flan-10k-flat/fix_punct | lora | | squad_v1_1_3_0_0 | phi-2 | sordonia/flan-10k-flat/squad_v1_1_3_0_0 | lora | | coqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/coqa_1_0_0 | lora | | glue_qnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_qnli_2_0_0 | lora | | wiki_qa_Jeopardy_style | phi-2 | sordonia/flan-10k-flat/wiki_qa_Jeopardy_style | lora | | qasc_qa_with_separated_facts_5 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_5 | lora | | glue_mnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mnli_2_0_0 | lora | | wiki_bio_key_content | phi-2 | sordonia/flan-10k-flat/wiki_bio_key_content | lora | | dream_generate_first_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_first_utterance | lora | | quartz_read_passage_below_choose | phi-2 | sordonia/flan-10k-flat/quartz_read_passage_below_choose | lora | | web_questions_question_answer | phi-2 | sordonia/flan-10k-flat/web_questions_question_answer | lora | | glue_stsb_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_stsb_2_0_0 | lora | | wmt16_translate_tr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_tr_en_1_0_0 | lora | | cot_qasc | phi-2 | sordonia/flan-10k-flat/cot_qasc | lora | | duorc_ParaphraseRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_title_generation | lora | | quail_description_context_question_answer_id | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_id | lora | | wiki_qa_Topic_Prediction_Question_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_Only | lora | | quoref_Find_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Find_Answer | lora | | social_i_qa_I_was_wondering | phi-2 | sordonia/flan-10k-flat/social_i_qa_I_was_wondering | lora | | wiki_hop_original_choose_best_object_affirmative_3 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_affirmative_3 | lora | | duorc_ParaphraseRC_build_story_around_qa | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_build_story_around_qa | lora | | qasc_qa_with_separated_facts_3 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_3 | lora | | race_middle_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_middle_Is_this_the_right_answer | lora | | paws_wiki_1_1_0 | phi-2 | sordonia/flan-10k-flat/paws_wiki_1_1_0 | lora | | app_reviews_categorize_rating_using_review | phi-2 | sordonia/flan-10k-flat/app_reviews_categorize_rating_using_review | lora | | anli_r3_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r3_0_1_0 | lora | | app_reviews_convert_to_rating | phi-2 | sordonia/flan-10k-flat/app_reviews_convert_to_rating | lora | | wiqa_what_is_the_final_step_of_the_following_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_is_the_final_step_of_the_following_process | lora | | adversarial_qa_droberta_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_answer_the_following_q | lora | | wiki_qa_Decide_good_answer | phi-2 | sordonia/flan-10k-flat/wiki_qa_Decide_good_answer | lora | | adversarial_qa_dbert_answer_the_following_q | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_answer_the_following_q | lora | | gem_dart_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_dart_1_1_0 | lora | | adversarial_qa_dbert_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_tell_what_it_is | lora | | quarel_choose_between | phi-2 | sordonia/flan-10k-flat/quarel_choose_between | lora | | duorc_ParaphraseRC_generate_question_by_answer | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question_by_answer | lora | | wiki_hop_original_generate_subject | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject | lora | | dream_baseline | phi-2 | sordonia/flan-10k-flat/dream_baseline | lora | | cos_e_v1_11_question_description_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_description_option_text | lora | | aeslc_1_0_0 | phi-2 | sordonia/flan-10k-flat/aeslc_1_0_0 | lora | | anli_r2_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r2_0_1_0 | lora | | dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_list_what_category_does_the_paragraph_belong_to | lora | | quail_context_question_description_answer_id | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_answer_id | lora | | race_middle_Select_the_best_answer_no_instructions_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_no_instructions_ | lora | | wmt16_translate_ro_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_ro_en_1_0_0 | lora | | race_high_Is_this_the_right_answer | phi-2 | sordonia/flan-10k-flat/race_high_Is_this_the_right_answer | lora | | quail_description_context_question_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_text | lora | | sciq_Direct_Question_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question_Closed_Book_ | lora | | openbookqa_0_1_0 | phi-2 | sordonia/flan-10k-flat/openbookqa_0_1_0 | lora | | duorc_SelfRC_title_generation | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_title_generation | lora | | cot_gsm8k | phi-2 | sordonia/flan-10k-flat/cot_gsm8k | lora | | quartz_answer_question_below | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_below | lora | | snli_1_1_0 | phi-2 | sordonia/flan-10k-flat/snli_1_1_0 | lora | | sciq_Multiple_Choice_Closed_Book_ | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Closed_Book_ | lora | | cot_strategyqa | phi-2 | sordonia/flan-10k-flat/cot_strategyqa | lora | | qasc_qa_with_separated_facts_4 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_separated_facts_4 | lora | | ropes_prompt_bottom_no_hint | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_no_hint | lora | | duorc_SelfRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_generate_question | lora | | quartz_given_the_fact_answer_the_q | phi-2 | sordonia/flan-10k-flat/quartz_given_the_fact_answer_the_q | lora | | anli_r1_0_1_0 | phi-2 | sordonia/flan-10k-flat/anli_r1_0_1_0 | lora | | wiki_qa_Topic_Prediction_Question_and_Answer_Pair | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Question_and_Answer_Pair | lora | | wiki_qa_Direct_Answer_to_Question | phi-2 | sordonia/flan-10k-flat/wiki_qa_Direct_Answer_to_Question | lora | | qasc_is_correct_2 | phi-2 | sordonia/flan-10k-flat/qasc_is_correct_2 | lora | | wiki_hop_original_generate_subject_and_object | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_generate_subject_and_object | lora | | ai2_arc_ARC_Challenge_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Challenge_1_0_0 | lora | | race_middle_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_middle_Select_the_best_answer_generate_span_ | lora | | quail_context_question_answer_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_answer_description_text | lora | | quail_context_question_description_text | phi-2 | sordonia/flan-10k-flat/quail_context_question_description_text | lora | | wiki_hop_original_choose_best_object_interrogative_2 | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_choose_best_object_interrogative_2 | lora | | duorc_SelfRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_movie_director | lora | | quoref_Given_Context_Answer_Question | phi-2 | sordonia/flan-10k-flat/quoref_Given_Context_Answer_Question | lora | | wiki_hop_original_explain_relation | phi-2 | sordonia/flan-10k-flat/wiki_hop_original_explain_relation | lora | | super_glue_record_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_record_1_0_2 | lora | | adversarial_qa_dbidaf_tell_what_it_is | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_tell_what_it_is | lora | | cot_ecqa_ii | phi-2 | sordonia/flan-10k-flat/cot_ecqa_ii | lora | | ropes_background_new_situation_answer | phi-2 | sordonia/flan-10k-flat/ropes_background_new_situation_answer | lora | | web_questions_short_general_knowledge_q | phi-2 | sordonia/flan-10k-flat/web_questions_short_general_knowledge_q | lora | | wiqa_what_might_be_the_first_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_first_step_of_the_process | lora | | duorc_SelfRC_answer_question | phi-2 | sordonia/flan-10k-flat/duorc_SelfRC_answer_question | lora | | ag_news_subset_1_0_0 | phi-2 | sordonia/flan-10k-flat/ag_news_subset_1_0_0 | lora | | race_middle_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_middle_Write_a_multi_choice_question_for_the_following_article | lora | | wmt14_translate_fr_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt14_translate_fr_en_1_0_0 | lora | | sciq_Direct_Question | phi-2 | sordonia/flan-10k-flat/sciq_Direct_Question | lora | | super_glue_multirc_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_multirc_1_0_2 | lora | | dbpedia_14_given_a_choice_of_categories_ | phi-2 | sordonia/flan-10k-flat/dbpedia_14_given_a_choice_of_categories_ | lora | | super_glue_wic_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wic_1_0_2 | lora | | social_i_qa_Show_choices_and_generate_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_answer | lora | | wiqa_what_might_be_the_last_step_of_the_process | phi-2 | sordonia/flan-10k-flat/wiqa_what_might_be_the_last_step_of_the_process | lora | | quoref_Answer_Question_Given_Context | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Question_Given_Context | lora | | quoref_Context_Contains_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Context_Contains_Answer | lora | | cos_e_v1_11_description_question_option_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_text | lora | | adversarial_qa_dbert_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_based_on | lora | | multi_news_1_0_0 | phi-2 | sordonia/flan-10k-flat/multi_news_1_0_0 | lora | | cos_e_v1_11_generate_explanation_given_text | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_generate_explanation_given_text | lora | | true_case | phi-2 | sordonia/flan-10k-flat/true_case | lora | | duorc_ParaphraseRC_movie_director | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_movie_director | lora | | quartz_answer_question_based_on | phi-2 | sordonia/flan-10k-flat/quartz_answer_question_based_on | lora | | bool_q_1_0_0 | phi-2 | sordonia/flan-10k-flat/bool_q_1_0_0 | lora | | quoref_Guess_Answer | phi-2 | sordonia/flan-10k-flat/quoref_Guess_Answer | lora | | quarel_do_not_use | phi-2 | sordonia/flan-10k-flat/quarel_do_not_use | lora | | cos_e_v1_11_explain_why_human | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_explain_why_human | lora | | wiki_qa_Generate_Question_from_Topic | phi-2 | sordonia/flan-10k-flat/wiki_qa_Generate_Question_from_Topic | lora | | kilt_tasks_hotpotqa_straighforward_qa | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_straighforward_qa | lora | | adversarial_qa_dbidaf_generate_question | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbidaf_generate_question | lora | | dbpedia_14_pick_one_category_for_the_following_text | phi-2 | sordonia/flan-10k-flat/dbpedia_14_pick_one_category_for_the_following_text | lora | | kilt_tasks_hotpotqa_final_exam | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_final_exam | lora | | quoref_Answer_Friend_Question | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Friend_Question | lora | | race_high_Write_a_multi_choice_question_for_the_following_article | phi-2 | sordonia/flan-10k-flat/race_high_Write_a_multi_choice_question_for_the_following_article | lora | | ropes_prompt_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_beginning | lora | | adversarial_qa_dbert_question_context_answer | phi-2 | sordonia/flan-10k-flat/adversarial_qa_dbert_question_context_answer | lora | | cot_creak | phi-2 | sordonia/flan-10k-flat/cot_creak | lora | | gem_e2e_nlg_1_1_0 | phi-2 | sordonia/flan-10k-flat/gem_e2e_nlg_1_1_0 | lora | | cos_e_v1_11_description_question_option_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_description_question_option_id | lora | | social_i_qa_Generate_the_question_from_the_answer | phi-2 | sordonia/flan-10k-flat/social_i_qa_Generate_the_question_from_the_answer | lora | | quarel_heres_a_story | phi-2 | sordonia/flan-10k-flat/quarel_heres_a_story | lora | | social_i_qa_Check_if_a_random_answer_is_valid_or_not | phi-2 | sordonia/flan-10k-flat/social_i_qa_Check_if_a_random_answer_is_valid_or_not | lora | | ropes_background_situation_middle | phi-2 | sordonia/flan-10k-flat/ropes_background_situation_middle | lora | | sciq_Multiple_Choice_Question_First | phi-2 | sordonia/flan-10k-flat/sciq_Multiple_Choice_Question_First | lora | | cot_strategyqa_ii | phi-2 | sordonia/flan-10k-flat/cot_strategyqa_ii | lora | | huggingface_xsum | phi-2 | sordonia/flan-10k-flat/huggingface_xsum | lora | | kilt_tasks_hotpotqa_complex_question | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_complex_question | lora | | wmt16_translate_fi_en_1_0_0 | phi-2 | sordonia/flan-10k-flat/wmt16_translate_fi_en_1_0_0 | lora | | ai2_arc_ARC_Easy_1_0_0 | phi-2 | sordonia/flan-10k-flat/ai2_arc_ARC_Easy_1_0_0 | lora | | stream_qed | phi-2 | sordonia/flan-10k-flat/stream_qed | lora | | definite_pronoun_resolution_1_1_0 | phi-2 | sordonia/flan-10k-flat/definite_pronoun_resolution_1_1_0 | lora | | super_glue_rte_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_rte_1_0_2 | lora | | ropes_new_situation_background_answer | phi-2 | sordonia/flan-10k-flat/ropes_new_situation_background_answer | lora | | dream_read_the_following_conversation_and_answer_the_question | phi-2 | sordonia/flan-10k-flat/dream_read_the_following_conversation_and_answer_the_question | lora | | cot_sensemaking | phi-2 | sordonia/flan-10k-flat/cot_sensemaking | lora | | wiki_qa_Topic_Prediction_Answer_Only | phi-2 | sordonia/flan-10k-flat/wiki_qa_Topic_Prediction_Answer_Only | lora | | duorc_ParaphraseRC_generate_question | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_generate_question | lora | | dream_generate_last_utterance | phi-2 | sordonia/flan-10k-flat/dream_generate_last_utterance | lora | | race_middle_Taking_a_test | phi-2 | sordonia/flan-10k-flat/race_middle_Taking_a_test | lora | | piqa_1_0_0 | phi-2 | sordonia/flan-10k-flat/piqa_1_0_0 | lora | | cot_ecqa | phi-2 | sordonia/flan-10k-flat/cot_ecqa | lora | | glue_mrpc_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_mrpc_2_0_0 | lora | | race_middle_Read_the_article_and_answer_the_question_no_option_ | phi-2 | sordonia/flan-10k-flat/race_middle_Read_the_article_and_answer_the_question_no_option_ | lora | | ropes_plain_background_situation | phi-2 | sordonia/flan-10k-flat/ropes_plain_background_situation | lora | | quail_description_context_question_answer_text | phi-2 | sordonia/flan-10k-flat/quail_description_context_question_answer_text | lora | | qasc_qa_with_combined_facts_1 | phi-2 | sordonia/flan-10k-flat/qasc_qa_with_combined_facts_1 | lora | | cot_creak_ii | phi-2 | sordonia/flan-10k-flat/cot_creak_ii | lora | | duorc_ParaphraseRC_decide_worth_it | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_decide_worth_it | lora | | quoref_Answer_Test | phi-2 | sordonia/flan-10k-flat/quoref_Answer_Test | lora | | wiki_bio_who | phi-2 | sordonia/flan-10k-flat/wiki_bio_who | lora | | kilt_tasks_hotpotqa_formulate | phi-2 | sordonia/flan-10k-flat/kilt_tasks_hotpotqa_formulate | lora | | glue_wnli_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_wnli_2_0_0 | lora | | gigaword_1_2_0 | phi-2 | sordonia/flan-10k-flat/gigaword_1_2_0 | lora | | quail_context_description_question_text | phi-2 | sordonia/flan-10k-flat/quail_context_description_question_text | lora | | dream_answer_to_dialogue | phi-2 | sordonia/flan-10k-flat/dream_answer_to_dialogue | lora | | cos_e_v1_11_question_option_description_id | phi-2 | sordonia/flan-10k-flat/cos_e_v1_11_question_option_description_id | lora | | duorc_ParaphraseRC_question_answering | phi-2 | sordonia/flan-10k-flat/duorc_ParaphraseRC_question_answering | lora | | wiki_qa_automatic_system | phi-2 | sordonia/flan-10k-flat/wiki_qa_automatic_system | lora | | adversarial_qa_droberta_based_on | phi-2 | sordonia/flan-10k-flat/adversarial_qa_droberta_based_on | lora | | super_glue_wsc_fixed_1_0_2 | phi-2 | sordonia/flan-10k-flat/super_glue_wsc_fixed_1_0_2 | lora | | word_segment | phi-2 | sordonia/flan-10k-flat/word_segment | lora | | quac_1_0_0 | phi-2 | sordonia/flan-10k-flat/quac_1_0_0 | lora | | quartz_paragraph_question_plain_concat | phi-2 | sordonia/flan-10k-flat/quartz_paragraph_question_plain_concat | lora | | wiqa_which_of_the_following_is_the_supposed_perturbation | phi-2 | sordonia/flan-10k-flat/wiqa_which_of_the_following_is_the_supposed_perturbation | lora | | quartz_use_info_from_question_paragraph | phi-2 | sordonia/flan-10k-flat/quartz_use_info_from_question_paragraph | lora | | ropes_plain_no_background | phi-2 | sordonia/flan-10k-flat/ropes_plain_no_background | lora | | race_high_Select_the_best_answer_generate_span_ | phi-2 | sordonia/flan-10k-flat/race_high_Select_the_best_answer_generate_span_ | lora | | glue_cola_2_0_0 | phi-2 | sordonia/flan-10k-flat/glue_cola_2_0_0 | lora | | social_i_qa_Show_choices_and_generate_index | phi-2 | sordonia/flan-10k-flat/social_i_qa_Show_choices_and_generate_index | lora | | ropes_prompt_bottom_hint_beginning | phi-2 | sordonia/flan-10k-flat/ropes_prompt_bottom_hint_beginning | lora | | stream_qed_ii | phi-2 | sordonia/flan-10k-flat/stream_qed_ii | lora | Last updated on: 2024-04-19 14:36:04+00:00
[ "SCIQ" ]
RWKV/v5-EagleX-v2-7B-pth
RWKV
null
[ "en", "dataset:cerebras/SlimPajama-627B", "dataset:EleutherAI/pile", "license:apache-2.0", "region:us" ]
"2024-04-17T04:48:41Z"
2024-05-02T22:02:31+00:00
0
0
--- datasets: - cerebras/SlimPajama-627B - EleutherAI/pile language: - en license: apache-2.0 --- ![An eagle flying high up in the sky](https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F304f2c7a-fc67-4df4-ba57-c6f38f86826c_2688x1536.png) ### RWKV EagleX 7B v2 Model > **!Important!: This is not meant to be used with huggingface transformers library** > [Use the Hugging Face varient instead, found here (v5-EagleX-v2-7B-HF)](https://huggingface.co/RWKV/v5-EagleX-v2-7B-HF) > > The following is the raw representation of the EagleX 7B v2 model. For use with our own set of trainers > > > This is not an instruct tune model! (soon...) ## Quickstart with the hugging face transformer library [See the huggingface version here (v5-EagleX-v2-7B-HF)](huggingface.co/RWKV/v5-EagleX-v2-7B-HF) ``` model = AutoModelForCausalLM.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True).to(torch.float32) tokenizer = AutoTokenizer.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True) ``` ## Evaluation The following shows the progression of the model from 1.1T trained to 2.25T trained. |Model |Eagle-7B-HF|EagleX-7B-HF-v1|EagleX-7B-HF-v2| |----------------------|-----------|---------------|---------------| |Param Count |7.52 B |7.52 B |7.52 B | |Tokens Trained |1.1 T |1.7 T |2.25 T | |avg_acc |0.4822 |0.5391 |0.5495 | |glue (acc) |0.5752 |0.7463 |0.7439 | |anli (acc) |0.3594 |0.4847 |0.5097 | |mnli (acc) |0.3802 |0.7928 |0.7884 | |mnli_mismatch (acc) |0.3687 |0.7985 |0.784 | |swag (acc) |0.568 |0.5814 |0.5905 | |lambada_standard (acc)|0.685 |0.686 |0.7004 | |lambada_openai (acc) |0.7425 |0.7522 |0.7502 | |mmlu (acc) |0.3321 |0.4014 |0.438 | |winogrande (acc) |0.674 |0.7206 |0.7332 | |wnli (acc) |0.4225 |0.4648 |0.493 | |truthfulqa (acc) |0.3303 |0.3268 |0.3401 | |logiqa (acc) |0.2458 |0.2458 |0.2458 | |logiqa2 (acc) |0.2494 |0.2595 |0.2621 | |sciq (acc) |0.955 |0.96 |0.93 | |piqa (acc) |0.7704 |0.7758 |0.7764 | |arc_easy (acc) |0.7382 |0.7555 |0.7445 | |arc_challenge (acc) |0.3951 |0.4087 |0.4155 | |hellaswag (acc) |0.5264 |0.5411 |0.56 | |openbookqa (acc) |0.302 |0.296 |0.304 | |mathqa (acc) |0.26 |0.26 |0.2593 | |arithmetic (acc) |0.245 |0.0634 |0.1703 | Compared against other top performing models in the same weight class. |Model |OLMo-7B |falcon-7b |Llama-2-7b-hf|EagleX-7B-HF-v2|Mistral-7B-v0.1| |----------------------|---------------|----------------|-------------|---------------|---------------| |Param Count |6.89 B |6.92 B |6.74 B |7.52 B |7.24 B | |Tokens Trained |2.5 T |1.5 T |2 T |2.25 T |2 - 7 T? | |avg_acc |0.4578 |0.4775 |0.5045 |0.5495 |0.5676 | |glue (acc) |0.474 |0.4578 |0.4289 |0.7439 |0.515 | |anli (acc) |0.3478 |0.3541 |0.3697 |0.5097 |0.3803 | |mnli (acc) |0.3294 |0.3893 |0.4269 |0.7884 |0.4542 | |mnli_mismatch (acc) |0.3348 |0.404 |0.4395 |0.784 |0.4632 | |swag (acc) |0.5512 |0.5685 |0.5658 |0.5905 |0.5756 | |lambada_standard (acc)|0.6396 |0.6868 |0.6808 |0.7004 |0.6944 | |lambada_openai (acc) |0.6872 |0.746 |0.7353 |0.7502 |0.7553 | |mmlu (acc) |0.2812 |0.2512 |0.4077 |0.438 |0.5964 | |winogrande (acc) |0.6725 |0.6709 |0.6914 |0.7332 |0.7364 | |wnli (acc) |0.5775 |0.4789 |0.4648 |0.493 |0.5775 | |truthfulqa (acc) |0.3015 |0.2826 |0.3205 |0.3401 |0.3537 | |logiqa (acc) |0.2335 |0.2151 |0.2535 |0.2458 |0.2427 | |logiqa2 (acc) |0.2506 |0.2252 |0.2564 |0.2621 |0.3022 | |sciq (acc) |0.927 |0.944 |0.939 |0.93 |0.959 | |piqa (acc) |0.7878 |0.7949 |0.7807 |0.7764 |0.8052 | |arc_easy (acc) |0.7353 |0.7479 |0.7643 |0.7445 |0.8081 | |arc_challenge (acc) |0.3677 |0.4027 |0.4309 |0.4155 |0.5009 | |hellaswag (acc) |0.5572 |0.5772 |0.5713 |0.56 |0.6131 | |openbookqa (acc) |0.292 |0.306 |0.316 |0.304 |0.33 | |mathqa (acc) |0.26 |0.2884 |0.2801 |0.2593 |0.3554 | |arithmetic (acc) |0.0069 |0.2367 |0.4703 |0.1703 |0.9004 | See the following, for the full details on this model: [https://blog.rwkv.com/p/eaglex-v2-soaring-past-llama2-7b](https://blog.rwkv.com/p/eaglex-v2-soaring-past-llama2-7b) ## Links - [Our wiki](https://wiki.rwkv.com) - [Full eval data](https://docs.google.com/spreadsheets/d/1CBLU6yKkW-8FMvGD4INO3qjeHZ0qkKnZFcM6n6lWNOs/edit#gid=912381775) - [Recursal.AI Cloud Platform](https://recursal.ai) - [HF Gradio Demo](https://huggingface.co/spaces/RWKV/v5-EagleX-v2-7B-gradio) - [Blog article, detailing our model launch](https://blog.rwkv.com/p/eaglex-v2-soaring-past-llama2-7b) ## Acknowledgement We are grateful for the help and support from the following key groups: - [Recursal.ai](https://recursal.ai) team for financing the GPU resources, and managing the training of this foundation model - you can run the Eagle line of RWKV models on their cloud / on-premise platform today. - EleutherAI for their support, especially in the v5/v6 Eagle/Finch paper - Linux Foundation AI & Data group for supporting and hosting the RWKV project
[ "SCIQ" ]
apple/OpenELM
apple
null
[ "arxiv:2404.14619", "license:other", "region:us" ]
"2024-04-17T20:01:04Z"
2024-05-02T00:54:23+00:00
0
1,429
--- license: other license_name: apple-sample-code-license license_link: LICENSE --- # OpenELM: An Efficient Language Model Family with Open Training and Inference Framework *Sachin Mehta, Mohammad Hossein Sekhavat, Qingqing Cao, Maxwell Horton, Yanzi Jin, Chenfan Sun, Iman Mirzadeh, Mahyar Najibi, Dmitry Belenko, Peter Zatloukal, Mohammad Rastegari* We introduce **OpenELM**, a family of **Open** **E**fficient **L**anguage **M**odels. OpenELM uses a layer-wise scaling strategy to efficiently allocate parameters within each layer of the transformer model, leading to enhanced accuracy. We pretrained OpenELM models using the [CoreNet](https://github.com/apple/corenet) library. We release both pretrained and instruction tuned models with 270M, 450M, 1.1B and 3B parameters. Our pre-training dataset contains RefinedWeb, deduplicated PILE, a subset of RedPajama, and a subset of Dolma v1.6, totaling approximately 1.8 trillion tokens. Please check license agreements and terms of these datasets before using them. See the list below for the details of each model: - [OpenELM-270M](https://huggingface.co/apple/OpenELM-270M) - [OpenELM-450M](https://huggingface.co/apple/OpenELM-450M) - [OpenELM-1_1B](https://huggingface.co/apple/OpenELM-1_1B) - [OpenELM-3B](https://huggingface.co/apple/OpenELM-3B) - [OpenELM-270M-Instruct](https://huggingface.co/apple/OpenELM-270M-Instruct) - [OpenELM-450M-Instruct](https://huggingface.co/apple/OpenELM-450M-Instruct) - [OpenELM-1_1B-Instruct](https://huggingface.co/apple/OpenELM-1_1B-Instruct) - [OpenELM-3B-Instruct](https://huggingface.co/apple/OpenELM-3B-Instruct) ```python from transformers import AutoModelForCausalLM openelm_270m = AutoModelForCausalLM.from_pretrained("apple/OpenELM-270M", trust_remote_code=True) openelm_450m = AutoModelForCausalLM.from_pretrained("apple/OpenELM-450M", trust_remote_code=True) openelm_1b = AutoModelForCausalLM.from_pretrained("apple/OpenELM-1_1B", trust_remote_code=True) openelm_3b = AutoModelForCausalLM.from_pretrained("apple/OpenELM-3B", trust_remote_code=True) openelm_270m_instruct = AutoModelForCausalLM.from_pretrained("apple/OpenELM-270M-Instruct", trust_remote_code=True) openelm_450m_instruct = AutoModelForCausalLM.from_pretrained("apple/OpenELM-450M-Instruct", trust_remote_code=True) openelm_1b_instruct = AutoModelForCausalLM.from_pretrained("apple/OpenELM-1_1B-Instruct", trust_remote_code=True) openelm_3b_instruct = AutoModelForCausalLM.from_pretrained("apple/OpenELM-3B-Instruct", trust_remote_code=True) ``` ## Usage We have provided an example function to generate output from OpenELM models loaded via [HuggingFace Hub](https://huggingface.co/docs/hub/) in `generate_openelm.py`. You can try the model by running the following command: ``` python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 ``` Please refer to [this link](https://huggingface.co/docs/hub/security-tokens) to obtain your hugging face access token. Additional arguments to the hugging face generate function can be passed via `generate_kwargs`. As an example, to speedup the inference, you can try [lookup token speculative generation](https://huggingface.co/docs/transformers/generation_strategies) by passing the `prompt_lookup_num_tokens` argument as follows: ``` python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 prompt_lookup_num_tokens=10 ``` Alternatively, try model-wise speculative generation with an [assistive model](https://huggingface.co/blog/assisted-generation) by passing a smaller model through the `assistant_model` argument, for example: ``` python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 --assistant_model [SMALLER_MODEL_NAME] ``` ## Main Results ### Zero-Shot | **Model Size** | **ARC-c** | **ARC-e** | **BoolQ** | **HellaSwag** | **PIQA** | **SciQ** | **WinoGrande** | **Average** | |-----------------------------------------------------------------------------|-----------|-----------|-----------|---------------|-----------|-----------|----------------|-------------| | [OpenELM-270M](https://huggingface.co/apple/OpenELM-270M) | 26.45 | 45.08 | **53.98** | 46.71 | 69.75 | **84.70** | **53.91** | 54.37 | | [OpenELM-270M-Instruct](https://huggingface.co/apple/OpenELM-270M-Instruct) | **30.55** | **46.68** | 48.56 | **52.07** | **70.78** | 84.40 | 52.72 | **55.11** | | [OpenELM-450M](https://huggingface.co/apple/OpenELM-450M) | 27.56 | 48.06 | 55.78 | 53.97 | 72.31 | 87.20 | 58.01 | 57.56 | | [OpenELM-450M-Instruct](https://huggingface.co/apple/OpenELM-450M-Instruct) | **30.38** | **50.00** | **60.37** | **59.34** | **72.63** | **88.00** | **58.96** | **59.95** | | [OpenELM-1_1B](https://huggingface.co/apple/OpenELM-1_1B) | 32.34 | **55.43** | 63.58 | 64.81 | **75.57** | **90.60** | 61.72 | 63.44 | | [OpenELM-1_1B-Instruct](https://huggingface.co/apple/OpenELM-1_1B-Instruct) | **37.97** | 52.23 | **70.00** | **71.20** | 75.03 | 89.30 | **62.75** | **65.50** | | [OpenELM-3B](https://huggingface.co/apple/OpenELM-3B) | 35.58 | 59.89 | 67.40 | 72.44 | 78.24 | **92.70** | 65.51 | 67.39 | | [OpenELM-3B-Instruct](https://huggingface.co/apple/OpenELM-3B-Instruct) | **39.42** | **61.74** | **68.17** | **76.36** | **79.00** | 92.50 | **66.85** | **69.15** | ### LLM360 | **Model Size** | **ARC-c** | **HellaSwag** | **MMLU** | **TruthfulQA** | **WinoGrande** | **Average** | |-----------------------------------------------------------------------------|-----------|---------------|-----------|----------------|----------------|-------------| | [OpenELM-270M](https://huggingface.co/apple/OpenELM-270M) | 27.65 | 47.15 | 25.72 | **39.24** | **53.83** | 38.72 | | [OpenELM-270M-Instruct](https://huggingface.co/apple/OpenELM-270M-Instruct) | **32.51** | **51.58** | **26.70** | 38.72 | 53.20 | **40.54** | | [OpenELM-450M](https://huggingface.co/apple/OpenELM-450M) | 30.20 | 53.86 | **26.01** | 40.18 | 57.22 | 41.50 | | [OpenELM-450M-Instruct](https://huggingface.co/apple/OpenELM-450M-Instruct) | **33.53** | **59.31** | 25.41 | **40.48** | **58.33** | **43.41** | | [OpenELM-1_1B](https://huggingface.co/apple/OpenELM-1_1B) | 36.69 | 65.71 | **27.05** | 36.98 | 63.22 | 45.93 | | [OpenELM-1_1B-Instruct](https://huggingface.co/apple/OpenELM-1_1B-Instruct) | **41.55** | **71.83** | 25.65 | **45.95** | **64.72** | **49.94** | | [OpenELM-3B](https://huggingface.co/apple/OpenELM-3B) | 42.24 | 73.28 | **26.76** | 34.98 | 67.25 | 48.90 | | [OpenELM-3B-Instruct](https://huggingface.co/apple/OpenELM-3B-Instruct) | **47.70** | **76.87** | 24.80 | **38.76** | **67.96** | **51.22** | ### OpenLLM Leaderboard | **Model Size** | **ARC-c** | **CrowS-Pairs** | **HellaSwag** | **MMLU** | **PIQA** | **RACE** | **TruthfulQA** | **WinoGrande** | **Average** | |-----------------------------------------------------------------------------|-----------|-----------------|---------------|-----------|-----------|-----------|----------------|----------------|-------------| | [OpenELM-270M](https://huggingface.co/apple/OpenELM-270M) | 27.65 | **66.79** | 47.15 | 25.72 | 69.75 | 30.91 | **39.24** | **53.83** | 45.13 | | [OpenELM-270M-Instruct](https://huggingface.co/apple/OpenELM-270M-Instruct) | **32.51** | 66.01 | **51.58** | **26.70** | **70.78** | 33.78 | 38.72 | 53.20 | **46.66** | | [OpenELM-450M](https://huggingface.co/apple/OpenELM-450M) | 30.20 | **68.63** | 53.86 | **26.01** | 72.31 | 33.11 | 40.18 | 57.22 | 47.69 | | [OpenELM-450M-Instruct](https://huggingface.co/apple/OpenELM-450M-Instruct) | **33.53** | 67.44 | **59.31** | 25.41 | **72.63** | **36.84** | **40.48** | **58.33** | **49.25** | | [OpenELM-1_1B](https://huggingface.co/apple/OpenELM-1_1B) | 36.69 | **71.74** | 65.71 | **27.05** | **75.57** | 36.46 | 36.98 | 63.22 | 51.68 | | [OpenELM-1_1B-Instruct](https://huggingface.co/apple/OpenELM-1_1B-Instruct) | **41.55** | 71.02 | **71.83** | 25.65 | 75.03 | **39.43** | **45.95** | **64.72** | **54.40** | | [OpenELM-3B](https://huggingface.co/apple/OpenELM-3B) | 42.24 | **73.29** | 73.28 | **26.76** | 78.24 | **38.76** | 34.98 | 67.25 | 54.35 | | [OpenELM-3B-Instruct](https://huggingface.co/apple/OpenELM-3B-Instruct) | **47.70** | 72.33 | **76.87** | 24.80 | **79.00** | 38.47 | **38.76** | **67.96** | **55.73** | See the technical report for more results and comparison. ## Evaluation ### Setup Install the following dependencies: ```bash # install public lm-eval-harness harness_repo="public-lm-eval-harness" git clone https://github.com/EleutherAI/lm-evaluation-harness ${harness_repo} cd ${harness_repo} # use main branch on 03-15-2024, SHA is dc90fec git checkout dc90fec pip install -e . cd .. # 66d6242 is the main branch on 2024-04-01 pip install datasets@git+https://github.com/huggingface/datasets.git@66d6242 pip install tokenizers>=0.15.2 transformers>=4.38.2 sentencepiece>=0.2.0 ``` ### Evaluate OpenELM ```bash # OpenELM-270M hf_model=apple/OpenELM-270M # this flag is needed because lm-eval-harness set add_bos_token to False by default, but OpenELM uses LLaMA tokenizer which requires add_bos_token to be True tokenizer=meta-llama/Llama-2-7b-hf add_bos_token=True batch_size=1 mkdir lm_eval_output shot=0 task=arc_challenge,arc_easy,boolq,hellaswag,piqa,race,winogrande,sciq,truthfulqa_mc2 lm_eval --model hf \ --model_args pretrained=${hf_model},trust_remote_code=True,add_bos_token=${add_bos_token},tokenizer=${tokenizer} \ --tasks ${task} \ --device cuda:0 \ --num_fewshot ${shot} \ --output_path ./lm_eval_output/${hf_model//\//_}_${task//,/_}-${shot}shot \ --batch_size ${batch_size} 2>&1 | tee ./lm_eval_output/eval-${hf_model//\//_}_${task//,/_}-${shot}shot.log shot=5 task=mmlu,winogrande lm_eval --model hf \ --model_args pretrained=${hf_model},trust_remote_code=True,add_bos_token=${add_bos_token},tokenizer=${tokenizer} \ --tasks ${task} \ --device cuda:0 \ --num_fewshot ${shot} \ --output_path ./lm_eval_output/${hf_model//\//_}_${task//,/_}-${shot}shot \ --batch_size ${batch_size} 2>&1 | tee ./lm_eval_output/eval-${hf_model//\//_}_${task//,/_}-${shot}shot.log shot=25 task=arc_challenge,crows_pairs_english lm_eval --model hf \ --model_args pretrained=${hf_model},trust_remote_code=True,add_bos_token=${add_bos_token},tokenizer=${tokenizer} \ --tasks ${task} \ --device cuda:0 \ --num_fewshot ${shot} \ --output_path ./lm_eval_output/${hf_model//\//_}_${task//,/_}-${shot}shot \ --batch_size ${batch_size} 2>&1 | tee ./lm_eval_output/eval-${hf_model//\//_}_${task//,/_}-${shot}shot.log shot=10 task=hellaswag lm_eval --model hf \ --model_args pretrained=${hf_model},trust_remote_code=True,add_bos_token=${add_bos_token},tokenizer=${tokenizer} \ --tasks ${task} \ --device cuda:0 \ --num_fewshot ${shot} \ --output_path ./lm_eval_output/${hf_model//\//_}_${task//,/_}-${shot}shot \ --batch_size ${batch_size} 2>&1 | tee ./lm_eval_output/eval-${hf_model//\//_}_${task//,/_}-${shot}shot.log ``` ## Bias, Risks, and Limitations The release of OpenELM models aims to empower and enrich the open research community by providing access to state-of-the-art language models. Trained on publicly available datasets, these models are made available without any safety guarantees. Consequently, there exists the possibility of these models producing outputs that are inaccurate, harmful, biased, or objectionable in response to user prompts. Thus, it is imperative for users and developers to undertake thorough safety testing and implement appropriate filtering mechanisms tailored to their specific requirements. ## Citation If you find our work useful, please cite: ```BibTex @article{mehtaOpenELMEfficientLanguage2024, title = {{OpenELM}: {An} {Efficient} {Language} {Model} {Family} with {Open} {Training} and {Inference} {Framework}}, shorttitle = {{OpenELM}}, url = {https://arxiv.org/abs/2404.14619v1}, language = {en}, urldate = {2024-04-24}, journal = {arXiv.org}, author = {Mehta, Sachin and Sekhavat, Mohammad Hossein and Cao, Qingqing and Horton, Maxwell and Jin, Yanzi and Sun, Chenfan and Mirzadeh, Iman and Najibi, Mahyar and Belenko, Dmitry and Zatloukal, Peter and Rastegari, Mohammad}, month = apr, year = {2024}, } @inproceedings{mehta2022cvnets, author = {Mehta, Sachin and Abdolhosseini, Farzad and Rastegari, Mohammad}, title = {CVNets: High Performance Library for Computer Vision}, year = {2022}, booktitle = {Proceedings of the 30th ACM International Conference on Multimedia}, series = {MM '22} } ```
[ "SCIQ" ]