AI & ML interests

Finetune Diffusion, Train Diffusion

AbhaykoulΒ 
posted an update 12 days ago
view post
Post
2939
πŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

🧠 What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think β†’ Answer β†’ Rethink β†’ Improve β†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
AbhaykoulΒ 
posted an update 26 days ago
view post
Post
2963
πŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

πŸš€ Key Features
- Intermediate thinking: Think β†’ Answer β†’ Rethink β†’ Answer β†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 1 reply
Β·
NymboΒ 
posted an update 27 days ago
view post
Post
2144
Anyone know how to reset Claude web's MCP config? I connected mine when the HF MCP first released with just the default example spaces added. I added lots of other MCP spaces but Claude.ai doesn't update the available tools... "Disconnecting" the HF integration does nothing, deleting it and adding it again does nothing.

Refreshing tools works fine in VS Code because I can manually restart it in mcp.json, but claude.ai has no such option. Anyone got any ideas?
Β·
AbhaykoulΒ 
posted an update about 1 month ago
view post
Post
4342
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
πŸ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of ∫0∞π‘₯3/π‘₯βˆ’1𝑑π‘₯ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
Β·
KingNishΒ 
posted an update about 2 months ago
view post
Post
972
What's currently the biggest gap in Open Source Datasets ??
Β·
NymboΒ 
posted an update 3 months ago
view post
Post
3655
Haven't seen this posted anywhere - Llama-3.3-8B-Instruct is available on the new Llama API. Is this a new model or did someone mislabel Llama-3.1-8B?
  • 1 reply
Β·
NymboΒ 
posted an update 3 months ago
view post
Post
2744
PSA for anyone using Nymbo/Nymbo_Theme or Nymbo/Nymbo_Theme_5 in a Gradio space ~

Both of these themes have been updated to fix some of the long-standing inconsistencies ever since the transition to Gradio v5. Textboxes are no longer bright green and in-line code is readable now! Both themes are now visually identical across versions.

If your space is already using one of these themes, you just need to restart your space to get the latest version. No code changes needed.
ehristoforuΒ 
posted an update 5 months ago
view post
Post
3572
Introducing our first standalone model – FluentlyLM Prinum

Introducing the first standalone model from Project Fluently LM! We worked on it for several months, used different approaches and eventually found the optimal one.

General characteristics:
- Model type: Causal language models (QwenForCausalLM, LM Transformer)
- Number of parameters: 32.5B
- Number of parameters (not embedded): 31.0B
- Number of layers: 64
- Context: 131,072 tokens
- Language(s) (NLP): English, French, Spanish, Russian, Chinese, Japanese, Persian (officially supported)
- License: MIT

Creation strategy:
The basis of the strategy is shown in Pic. 2.
We used Axolotl & Unsloth for SFT-finetuning with PEFT LoRA (rank=64, alpha=64) and Mergekit for SLERP and TIES mergers.

Evolution:
πŸ† 12th place in the Open LLM Leaderboard ( open-llm-leaderboard/open_llm_leaderboard) (21.02.2025)

Detailed results and comparisons are presented in Pic. 3.

Links:
- Model: fluently-lm/FluentlyLM-Prinum
- GGUF version: mradermacher/FluentlyLM-Prinum-GGUF
- Demo on ZeroGPU: ehristoforu/FluentlyLM-Prinum-demo
  • 7 replies
Β·
ameerazam08Β 
posted an update 6 months ago
AbhaykoulΒ 
posted an update 6 months ago
view post
Post
4773
πŸ”₯ THE WAIT IS OVER... HAI-SER IS HERE! πŸ”₯

Yo fam, this ain't just another AI dropβ€” this is the FUTURE of emotional intelligence! πŸš€

Introducing HAI-SER, powered by Structured Emotional Reasoning (SER), the next-level AI that doesn’t just understand your wordsβ€”it feels you, analyzes your emotions, and helps you navigate life’s toughest moments. πŸ’‘

πŸ’₯ What makes HAI-SER a game-changer?
πŸ”Ή Emotional Vibe Check – Gets the mood, energy, and what’s really going on 🎭
πŸ”Ή Mind-State Analysis – Breaks down your thoughts, beliefs, and patterns 🀯
πŸ”Ή Root Cause Deep-Dive – Unpacks the WHY behind your emotions πŸ’‘
πŸ”Ή Impact Check – Sees how it’s affecting your life and mental health πŸ’”
πŸ”Ή Safety Check – Prioritizes your well-being and crisis management 🚨
πŸ”Ή Healing Game Plan – Custom strategies to help you bounce back πŸ’ͺ
πŸ”Ή Growth Potential – Turns struggles into opportunities for self-improvement πŸ“ˆ
πŸ”Ή How to Approach – Teaches you and others how to communicate and heal 🀝
πŸ”Ή Personalized Response – Not just generic adviceβ€”real talk, tailored to YOU πŸ’―

No more robotic AI responses. No more surface-level advice. HAI-SER gets deep, analyzing emotions with precision and giving real, actionable support.

This ain’t just AIβ€”this is your digital therapist, life coach, and hype squad all in one. Whether it’s mental health, career struggles, relationships, or personal growth, HAI-SER has your back.

πŸš€ The future of emotionally intelligent AI is HERE.
Are you ready? πŸ”₯πŸ’―

HelpingAI/HAI-SER
Β·
1aurentΒ 
posted an update 7 months ago
ehristoforuΒ 
posted an update 7 months ago
view post
Post
4409
βœ’οΈ Ultraset - all-in-one dataset for SFT training in Alpaca format.
fluently-sets/ultraset

❓ Ultraset is a comprehensive dataset for training Large Language Models (LLMs) using the SFT (instruction-based Fine-Tuning) method. This dataset consists of over 785 thousand entries in eight languages, including English, Russian, French, Italian, Spanish, German, Chinese, and Korean.

🀯 Ultraset solves the problem faced by users when selecting an appropriate dataset for LLM training. It combines various types of data required to enhance the model's skills in areas such as text writing and editing, mathematics, coding, biology, medicine, finance, and multilingualism.

πŸ€— For effective use of the dataset, it is recommended to utilize only the "instruction," "input," and "output" columns and train the model for 1-3 epochs. The dataset does not include DPO or Instruct data, making it suitable for training various types of LLM models.

❇️ Ultraset is an excellent tool to improve your language model's skills in diverse knowledge areas.
AbhaykoulΒ 
posted an update 7 months ago
view post
Post
2204
πŸ”₯ BIG ANNOUNCEMENT: THE HELPINGAI API IS LIVE! πŸ”₯

Yo, the moment you’ve all been waiting for is here! πŸš€ The HelpingAI API is now LIVE and ready to level up your projects! πŸ”₯ We’re bringing that next-level AI goodness straight to your fingertips. πŸ’―

No more waitingβ€” it’s time to build something epic! πŸ™Œ

From now on, you can integrate our cutting-edge AI models into your own applications, workflows, and everything in between. Whether you’re a developer, a creator, or just someone looking to make some serious moves, this is your chance to unlock the full potential of emotional intelligence and adaptive AI.

Check out the docs πŸ”₯ and let’s get to work! πŸš€

πŸ‘‰ Check out the docs and start building (https://helpingai.co/docs)
πŸ‘‰ Visit the HelpingAI website (https://helpingai.co/)
Β·
KingNishΒ 
posted an update 10 months ago
KingNishΒ 
posted an update 10 months ago
view post
Post
8288
Exciting news! Introducing super-fast AI video assistant, currently in beta. With a minimum latency of under 500ms and an average latency of just 600ms.

DEMO LINK:
KingNish/Live-Video-Chat
  • 1 reply
Β·
KingNishΒ 
posted an update 10 months ago
KingNishΒ 
posted an update 10 months ago
view post
Post
3612
Mistral Nemo is better than many models in 1st grader level reasoning.
KingNishΒ 
posted an update 11 months ago
view post
Post
3943
I am experimenting with Flux and trying to push it to its limits without training (as I am GPU-poor πŸ˜…).
I found some flaws in the pipelines, which I resolved, and now I am able to generate an approx similar quality image as Flux Schnell 4 steps in just 1 step.
Demo Link:
KingNish/Realtime-FLUX

  • 1 reply
Β·
KingNishΒ 
posted an update 11 months ago
view post
Post
1932
I am excited to announce a major speed updated in Voicee, a superfast voice assistant.

It has now achieved latency <250 ms.
While its average latency is about 500ms.
KingNish/Voicee

This become Possible due to newly launched @sambanovasystems cloud.

You can also use your own API Key to get fastest speed.
You can get on from here: https://cloud.sambanova.ai/apis

For optimal performance use Google Chrome.

Please try Voicee and share your valuable feedback to help me further improve its performance and usability.
Thank you!
1aurentΒ 
posted an update 11 months ago
view post
Post
1399
Hey everyone πŸ€—!
We (finegrain) have created some custom ComfyUI nodes to use our refiners micro-framework inside comfy! πŸŽ‰

We only support our new Box Segmenter at the moment, but we're thinking of adding more nodes since there seems to be a demand for it. We leverage the new (beta) Comfy Registry to host our nodes. They are available at: https://registry.comfy.org/publishers/finegrain/nodes/comfyui-refiners. You can install them by running:
comfy node registry-install comfyui-refiners

Or by unzipping the archive you can download by clicking "Download Latest" into your custom_nodes comfy folder.
We are eager to hear your feedbacks and suggestions for new nodes and how you'll use them! πŸ™