AI & ML interests

None defined yet.

Recent Activity

AnSungJae3489ย 
posted an update 7 months ago
view post
Post
2632
ShareGPT? How about ShareGPT-X?

We release **92K** Human with LLM conversations as a refresh and update over the original ShareGPT Dataset.

DSULT-Core/ShareGPT-X
Abhaykoulย 
posted an update 7 months ago
view post
Post
3339
๐Ÿš€ Ever dreamed of training your own Large Language Model from scratch? What if I told you it doesn't require a supercomputer or PhD in ML? ๐Ÿคฏ

Introducing LLM Trainer - the educational framework that makes LLM training accessible to EVERYONE! Whether you're on a CPU-only laptop or scaling to distributed GPUs, we've got you covered. ๐Ÿ’ปโžก๏ธ๐Ÿ–ฅ๏ธ

Why LLM Trainer? Because existing tools are either too simplistic (hiding the magic) or too complex (requiring expert knowledge). We bridge the gap with:

๐ŸŽ“ Educational transparency - every component built from scratch with clear code
๐Ÿ’ป CPU-first approach - start training immediately, no GPU needed
๐Ÿ”ง Full customization - modify anything you want
๐Ÿ“ˆ Seamless scaling - from laptop to cluster without code changes
๐Ÿค HuggingFace integration - works with existing models & tokenizers

Key highlights:
โœ… Built-in tokenizers (BPE, WordPiece, HF wrappers)
โœ… Complete Transformer implementation from scratch
โœ… Optimized for CPU training
โœ… Advanced features: mixed precision, gradient checkpointing, multiple generation strategies
โœ… Comprehensive monitoring & metrics

Perfect for:
- Students learning transformers
- Researchers prototyping new ideas
- Developers building domain-specific models

Ready to train your first LLM? It's easier than you think!

๐Ÿ”— Check it out: https://github.com/HelpingAI/llm-trainer
๐Ÿ“š Docs: Getting Started Guide
๐Ÿ’ฌ Join the community: GitHub Discussions

#AI #MachineLearning #LLM #DeepLearning #OpenSource #Python #HuggingFace #NLP

Special thanks to HuggingFace and PyTorch teams for the amazing ecosystem! ๐Ÿ™
  • 1 reply
ยท
Abhaykoulย 
posted an update 9 months ago
view post
Post
4173
๐Ÿš€ Dhanishtha-2.0-preview-0825 Is Here

The Intermediate Thinking Model just leveled up again.

With sharper reasoning, better tool use, and expanded capabilities, Dhanishtha-2.0-preview-0825 is now live and ready to impress.

๐Ÿง  What Makes Dhanishtha Special?
Unlike typical CoT models that only thinks one time, Dhanishtha thinks iteratively:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

๐Ÿ”— Try it now: HelpingAI/Dhanishtha-2.0-preview-0825

๐Ÿ”ž Dhanishtha NSFW Preview

For those exploring more expressive and immersive roleplay scenarios, weโ€™re also releasing:

HelpingAI/Dhanishtha-nsfw
A specialized version tuned for adult-themed interactions and character-driven roleplay.

๐Ÿ”— Explore it here: HelpingAI/Dhanishtha-nsfw

๐Ÿ’ฌ You can also try all of these live at chat.helpingai.co
  • 4 replies
ยท
Abhaykoulย 
posted an update 9 months ago
view post
Post
3167
๐ŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

๐Ÿง  What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
Abhaykoulย 
posted an update 10 months ago
view post
Post
3125
๐ŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

๐Ÿš€ Key Features
- Intermediate thinking: Think โ†’ Answer โ†’ Rethink โ†’ Answer โ†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 2 replies
ยท
Abhaykoulย 
posted an update 10 months ago
view post
Post
4688
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
๐Ÿ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of โˆซ0โˆž๐‘ฅ3/๐‘ฅโˆ’1๐‘‘๐‘ฅ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
ยท