Some One

s0me-0ne

AI & ML interests

None yet

Recent Activity

Organizations

None yet

s0me-0ne's activity

reacted to AdinaY's post with ๐Ÿ‘€ 1 day ago
view post
Post
1700
AutoGLM ๆฒ‰ๆ€๐Ÿ’ซ FREE AI Agent released by ZhipuAI

โœจ Think & Act simultaneously
โœจ Based on a fully self-developed stack: GLM-4 for general, GLM-Z1 for inference, and GLM-Z1-Rumination for rumination
โœจ Will openly share these models on April 14 ๐Ÿคฏ

Preview version๐Ÿ‘‰ https://autoglm-research.zhipuai.cn/?channel=autoglm_android
reacted to prithivMLmods's post with ๐Ÿ”ฅ 12 days ago
view post
Post
2255
Play with Orpheus TTS, a Llama-based Speech-LLM designed for high-quality, empathetic text-to-speech generation. This model has been fine-tuned to deliver human-level speech synthesis ๐Ÿ”ฅ๐Ÿ—ฃ๏ธ

๐Ÿ‘‰GitHub [ Demo ] : https://github.com/PRITHIVSAKTHIUR/Orpheus-TTS-Edge

Demo supporting both text-to-speech and text-to-llm responses in speech.

> voice: tara, dan, emma, josh
> emotion: <laugh>, <chuckle>, <sigh>, <cough>, <sniffle>, <groan>, <yawn>, <gasp>.

๐Ÿฅ Orpheus-3b-0.1-ft
Model Page: canopylabs/orpheus-3b-0.1-ft

๐Ÿฅ Orpheus-3b-0.1-ft
Colab Inference Notebook: https://colab.research.google.com/drive/1KhXT56UePPUHhqitJNUxq63k-pQomz3N?usp=sharing

๐Ÿฅ Finetune [ orpheus-3b-0.1-pretrained ]
Resource: https://github.com/canopyai/Orpheus-TTS/tree/main/finetune

๐Ÿฅ Model-releases:
https://canopylabs.ai/model-releases
  • 1 reply
ยท
upvoted an article about 2 months ago
view article
Article

Open-source DeepResearch โ€“ Freeing our search agents

โ€ข 1.2k
reacted to fdaudens's post with ๐Ÿ”ฅ about 2 months ago
view post
Post
3493
๐ŸŽฏ Kokoro TTS just hit v1.0! ๐Ÿš€

Small but mighty: 82M parameters, runs locally, speaks multiple languages. The best part? It's Apache 2.0 licensed!
This could unlock so many possibilities โœจ

Check it out: hexgrad/Kokoro-82M
  • 1 reply
ยท
reacted to hexgrad's post with ๐Ÿ”ฅ 2 months ago
reacted to m-ric's post with ๐Ÿ”ฅ 2 months ago
view post
Post
3363
Today we make the biggest release in smolagents so far: ๐˜„๐—ฒ ๐—ฒ๐—ป๐—ฎ๐—ฏ๐—น๐—ฒ ๐˜ƒ๐—ถ๐˜€๐—ถ๐—ผ๐—ป ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€, ๐˜„๐—ต๐—ถ๐—ฐ๐—ต ๐—ฎ๐—น๐—น๐—ผ๐˜„๐˜€ ๐˜๐—ผ ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐—ฝ๐—ผ๐˜„๐—ฒ๐—ฟ๐—ณ๐˜‚๐—น ๐˜„๐—ฒ๐—ฏ ๐—ฏ๐—ฟ๐—ผ๐˜„๐˜€๐—ถ๐—ป๐—ด ๐—ฎ๐—ด๐—ฒ๐—ป๐˜๐˜€! ๐Ÿฅณ

Our agents can now casually open up a web browser, and navigate on it by scrolling, clicking elements on the webpage, going back, just like a user would.

The demo below shows Claude-3.5-Sonnet browsing GitHub for task: "Find how many commits the author of the current top trending repo did over last year."
Hi @mlabonne !

Go try it out, it's the most cracked agentic stuff I've seen in a while ๐Ÿคฏ (well, along with OpenAI's Operator who beat us by one day)

For more detail, read our announcement blog ๐Ÿ‘‰ https://huggingface.co/blog/smolagents-can-see
The code for the web browser example is here ๐Ÿ‘‰ https://github.com/huggingface/smolagents/blob/main/examples/vlm_web_browser.py
ยท