AI & ML interests

Training a Traditional Chinese Large Language Model.

prithivMLmodsย 
posted an update 3 days ago
view post
Post
669
GLM OCR, a multimodal OCR model for complex document understanding, built on the GLM-V encoderโ€“decoder architecture. It delivers high accuracy and strong generalization with a blazing-fast inference pipeline. The demo is live . Try it now. ๐Ÿค—๐Ÿš€

โœจ Demo: prithivMLmods/GLM-OCR-Demo
โœจ Multimodal Implementations: https://huggingface.co/collections/prithivMLmods/multimodal-implementations
โœจ GitHub: https://github.com/PRITHIVSAKTHIUR/GLM-OCR-Demo
Sri-Vigneshwar-DJย 
posted an update 4 days ago
view post
Post
1338
Just released a new dataset designed for training reasoning models on Meta (Facebook/Instagram) advertising fatigue detection!

What is it? A GRPO (Group Relative Policy Optimization) training dataset with 200+ carefully crafted scenarios covering:

๐Ÿ” Fatigue Signal Detection: CTR drops, CPM spikes, frequency analysis
๐Ÿฉบ Performance Diagnosis: Root cause analysis frameworks
๐Ÿ“‹ Strategy: Creative refresh cadence, testing frameworks
๐Ÿ“Š Analysis: ROI calculations, metric interpretation
Why GRPO? GRPO training helps models learn structured reasoning. Each response follows the <thinking> and <answer> format.

Check it out here: Sri-Vigneshwar-DJ/meta-fatigue-grpo-dataset
prithivMLmodsย 
posted an update 4 days ago
view post
Post
2077
Introducing the Qwen-Image-Edit-3D-Lighting-Control app, featuring 8ร— horizontal and 3ร— elevational lighting positions for precise 3D lighting control. It enables studio-level lighting using fast Qwen Image Edit fast inference, paired with Multi-Angle-Lighting adapters. ๐Ÿ”ฆ

๐Ÿ”ฅ Space: prithivMLmods/Qwen-Image-Edit-3D-Lighting-Control
โœ… Collection: https://huggingface.co/collections/prithivMLmods/image-generation-apps-collection
๐Ÿ“‚ GitHub: https://github.com/PRITHIVSAKTHIUR/Qwen-Image-Edit-3D-Lighting-Control
prithivMLmodsย 
posted an update 10 days ago
view post
Post
3586
Daggr UI version of the Qwen3-TTS demo.๐Ÿ”ฅ
(custom voice, voice design, qwen3-asr and voice cloning) nodes.
No remote spaces used for API inference; all functions run in-app fn.
Powered by t4-m and built with daggr@0.5.2 and gradio@6.

๐Ÿ‘‰Demo: prithivMLmods/Qwen3-TTS-Daggr-UI
โญGithub: https://github.com/PRITHIVSAKTHIUR/Qwen3-TTS-Daggr-UI
  • 1 reply
ยท
prithivMLmodsย 
posted an update 13 days ago
view post
Post
2664
Qwen-Image-Edit-Object-Manipulator Space is now featured in Hugging Face Space of the Week. It enables object manipulation such as extracting objects, adding designs, and removing objects or designs from the red highlighted area using specialized adapters.

๐Ÿ”ฅDo enjoy the demo! ~ prithivMLmods/Qwen-Image-Edit-Object-Manipulator

Collections:
๐ŸงจAdapters-1: https://huggingface.co/collections/prithivMLmods/qwen-image-edit-exps
๐ŸงจAdapters-2: https://huggingface.co/collections/prithivMLmods/qie-jan-23-26
๐ŸงจAdapters-3: https://huggingface.co/collections/prithivMLmods/qwen-image-edit-object-manipulator

โญGithub: https://github.com/PRITHIVSAKTHIUR/Qwen-Image-Edit-Object-Manipulator

To learn more, visit the app page or the respective model pages.
Sri-Vigneshwar-DJย 
posted an update 13 days ago
view post
Post
196
๐Ÿ™๏ธ Hugging Face Community Post
Title: ๐Ÿงฌ Experimenting with "Dynamic Chaos" in Tamil SLMs

Hi everyone! I just published a new experimental study on Small Language Model (SLM) resilience.

I took the Qwen2.5-0.5B model and put it through a "Chaos Phase" to see how much weight data a tiny model can lose before its understanding of classical Tamil grammar breaks.

Key highlights of the study:

Target Data: Fine-tuned on the Thirukkural (1,330 couplets + modern explanations).
The Chaos Step: Applied 20% random weight pruning but implemented "Layer Protection" for the Token Embeddings and LM Head to keep the characters readable.
Compression: 4-bit (Q4_K_M) quantization for extreme efficiency.
Result: A surrealist classical Tamil model that is ultra-light (~300MB) and ultra-fast!

Check out the model and the experiment logic here: Sri-Vigneshwar-DJ/qwen-tamil-chaos-v1
prithivMLmodsย 
posted an update 16 days ago
view post
Post
3023
Introducing QIE-2511-Zoom-Master for highlight-guided area zoom-in, enabling lossless zooming within a drawn square area, and QIE-2511-Object-Remover-v2 for precise object or highlight-guided area cleanup. These experimental adapters are trained based on QIE-2511. Find the adapters below.

๐Ÿ•น๏ธQIE-2511-Zoom-Master : prithivMLmods/QIE-2511-Zoom-Master
๐Ÿ•น๏ธQIE-2511-Object-Remover-v2: prithivMLmods/QIE-2511-Object-Remover-v2

๐Ÿค—Demo: prithivMLmods/Qwen-Image-Edit-Object-Manipulator

๐Ÿ“‚Collection: https://huggingface.co/collections/prithivMLmods/qwen-image-edit-exps

To learn more, visit the app page or the respective model pages.
  • 2 replies
ยท
Sri-Vigneshwar-DJย 
posted an update 21 days ago
view post
Post
307
Performance Marketing meets "Thinking Mode" ๐Ÿง 

Iโ€™m excited to release hawky-ai-Qwen3-0.6B-Marketing-MoT, a specialized SLM designed for deep strategic reasoning in performance marketing.

While small at 0.6B parameters, this model punches way above its weight class by utilizing a Mixture of Thoughts (MoT) framework. It doesn't just give you an answer; it thinks through the logic of Meta Ads scaling, GA4 attribution, and unit economics before providing a strategic recommendation.

Key Features:

Thinking-First: Trained on 1,500+ critical thinking scenarios.
MoT Framework: 5 distinct reasoning styles (Linear, Exploratory, Critical, Deconstructive, Analogical).
SLM Speed: Perfect for low-latency, high-precision marketing audits.
Check it out on Hugging Face: ๐Ÿ”— Sri-Vigneshwar-DJ/hawky-ai-Qwen3-0.6B-Marketing-MoT
prithivMLmodsย 
posted an update 28 days ago
view post
Post
5552
LTX-2 Camera-Control LoRA demo with dolly-in/out and dolly-left/right is now available on Hugging Face, paired with ltx-2-19b-distilled-lora for fast inference. It also includes dynamic GPU duration adjustments for long video generations. Click the related Space links below.

๐Ÿค—Try it now on : prithivMLmods/LTX-2-LoRAs-Camera-Control-Dolly
โญGithub: https://github.com/PRITHIVSAKTHIUR/LTX-2-LoRAs-Camera-Control-Dolly
๐Ÿ•น๏ธCollection: https://huggingface.co/collections/prithivMLmods/image-generation-apps-collection

To learn more, visit the app page or the respective model pages.
  • 2 replies
ยท
Sri-Vigneshwar-DJย 
posted an update 28 days ago
view post
Post
2182
Introducing Hawky-AI H1 4B PM: The First Open-Source LLM for Performance Marketing ๐ŸŽฏ

Hey HF Community! ๐Ÿ‘‹

Just released the first LLM fine-tuned specifically for Performance Marketing.
What is it?
Gemma 3 4B distilled from Claude Opus 4.5 with expert-level marketing knowledge.
Covers:
๐Ÿ“ฑ Meta Ads (campaign structure, bidding, scaling, creative fatigue)
๐Ÿ” Google Ads (Quality Score, Performance Max, lead gen)
๐Ÿ“Š Measurement (ROAS vs MER, incrementality, LTV:CAC)
๐ŸŽจ Creative Strategy (hook rates, A/B testing, funnel creative)
Why we built it:
Generic LLMs say "optimize your targeting" โ€” not helpful. This model gives specific frameworks like "frequency at 4.5 + CTR drop = creative fatigue, here's the fix..."
Technical:

Base: Gemma 3 4B
Method: QLoRA (r=64)
Teacher: Claude Opus 4.5

๐Ÿ”— Model: Sri-Vigneshwar-DJ/hawky-ai-H1-4b-PM
Built by Hawky.ai

Try it and let us know what you think! ๐Ÿš€
Sri-Vigneshwar-DJย 
posted an update about 1 month ago
view post
Post
1382
๐Ÿฆ… Introducing Hawky AI H1 Mini 4B: A Domain-Specific Model for Performance Marketing

Hey HuggingFace community! ๐Ÿ‘‹

We're excited to share our first open-source release: **Hawky AI H1 Mini 4B Experimental** - a Gemma 3 4B model fine-tuned specifically for Meta advertising and performance marketing strategy.

๐ŸŽฏ Why We Built This

At [Hawky.ai](https://hawky.ai), we build AI-powered creative intelligence tools for performance marketers. We work with major agencies (WPP, Madison, GroupM) and brands (TVS Motors, Tanishq, Bajaj Finserv) on campaign optimization.

We wanted to explore: Can a small, domain-specific model provide expert-level guidance on performance marketing?

Specifically, we focused on Meta's Andromeda algorithm - the AI system that now powers ad delivery across Facebook and Instagram. Understanding Andromeda is crucial for modern media buying, but the knowledge is scattered and constantly evolving.

๐Ÿง  What Makes This Different

Chain-of-Thought Reasoning
The model doesn't just answer - it **thinks through problems** step-by-step:

Sri-Vigneshwar-DJ/hawky-ai-h1-mini-4b-experimental
prithivMLmodsย 
posted an update about 1 month ago
view post
Post
2466
Dropping Image Edit (Object Manipulator): Add or remove specified objects/designs, with flexible support for both single-image and multi-image modes.

๐Ÿค— Demo: prithivMLmods/Qwen-Image-Edit-Object-Manipulator

Qwen-Image-Edit-2511-Object-Remover is an adapter (LoRA) developed for Qwenโ€™s Qwen-Image-Edit-2511 image-to-image model. It is specifically designed for precise object removal from images.

โญ Model: prithivMLmods/Qwen-Image-Edit-2511-Object-Remover

Qwen-Image-Edit-2511-Object-Adder is an adapter (LoRA) developed for Qwenโ€™s Qwen-Image-Edit-2511 image-to-image model. It is specifically designed for precise object addition to images.

โญ Model: prithivMLmods/Qwen-Image-Edit-2511-Object-Adder

๐Ÿ•น๏ธ Collection: https://huggingface.co/collections/prithivMLmods/qwen-image-edit-object-manipulator
๐Ÿ•น๏ธ github: https://github.com/PRITHIVSAKTHIUR/Qwen-Image-Edit-Object-Manipulator

To learn more, visit the app page or the respective model pages.
Sri-Vigneshwar-DJย 
posted an update about 1 month ago
view post
Post
935
Domain-specific reasoning is crucial when working with big-budget campaigns on Meta. That's why we've launched an experimental Chain-of-Thought (CoT) reasoning model for critical thinking, tailored to Meta's Andromeda algorithm-based campaign structuring and optimization.

Sri-Vigneshwar-DJ/hawky-ai-h1-mini-1b-experimental
Sri-Vigneshwar-DJย 
posted an update about 1 month ago
view post
Post
2958
The recent update to Meta's ad algorithm is very difficult to crack, and even the latest models struggle to keep up with it. To address this, we've created a small experimental dataset for fine-tuning models to better tackle Meta's Andromeda algorithm: Sri-Vigneshwar-DJ/hawky-ai-andromeda-dataset
Sri-Vigneshwar-DJย 
posted an update about 1 month ago
prithivMLmodsย 
posted an update about 1 month ago
view post
Post
4211
Update: TRELLIS.2 (Text to 3D, Image to 3D) Gradio with Rerun Embedded demo with improved visualization of the 3D model previewer is now available on Hugging Face. Generate assets and view them in the 3D viewer, powered and streamlined with Microsoftโ€™s TRELLIS.2 and Tongyi-MAIโ€™s Z-Image-Turbo models.

๐Ÿค— TRELLIS.2 (Demo): prithivMLmods/TRELLIS.2-Text-to-3D
๐Ÿ•น๏ธ GitHub: https://github.com/PRITHIVSAKTHIUR/TRELLIS.2-Text-to-3D-RERUN
๐Ÿ•น๏ธ Collection: https://huggingface.co/collections/prithivMLmods/multimodal-implementations

To know more about it, visit the app page or the respective model page!
prithivMLmodsย 
posted an update about 1 month ago
view post
Post
4267
Introducing the Qwen-Image-Edit-2511-LoRAs-Fast demo, featuring image property comparison and contrast, built on top of Gradio and the combined Rerun SDK. It supports single and multi-image edits with existing LoRAs that are lazily loaded. (Note: This is still an experimental Space for Qwen-Image-Edit-2511.)

โญ Space Demo: prithivMLmods/Qwen-Image-Edit-2511-LoRAs-Fast
โญ GitHub: https://github.com/PRITHIVSAKTHIUR/Qwen-Image-Edit-2511-LoRAs-Fast-Multi-Image-Rerun
โญ Collection: https://huggingface.co/collections/prithivMLmods/image-generation-apps-collection

To know more about it, visit the app page or the respective model page!
  • 2 replies
ยท
prithivMLmodsย 
posted an update about 2 months ago
view post
Post
3738
Introducing demos for new SOTA models from AI2: SAGE-MM (Smart Any-Horizon Agents for Long-Video Reasoning) and Molmo-2, an open vision-language model that supports multi-image (QA and pointing) and video (QA, pointing, and tracking). The respective demo-related collections are listed below. ๐ŸŽƒ๐Ÿ”ฅ

โœจ SAGE-MM [Video-Reasoning]: prithivMLmods/SAGE-MM-Video-Reasoning
โœจ Molmo2 [Demo]: prithivMLmods/Molmo2-HF-Demo

๐ŸŽƒ GitHub[SAGE-MM]: https://github.com/PRITHIVSAKTHIUR/SAGE-MM-Video-Reasoning
๐ŸŽƒ GitHub[Molmo2]: https://github.com/PRITHIVSAKTHIUR/Molmo2-HF-Demo
๐ŸŽƒ Multimodal Implementations: https://huggingface.co/collections/prithivMLmods/multimodal-implementations

To know more about it, visit the app page or the respective model page!
  • 1 reply
ยท
prithivMLmodsย 
posted an update about 2 months ago
view post
Post
2105
Introducing TRELLIS.2 Text-to-3D. The demo for the TRELLIS.2-4B (Image-to-3D) model is streamlined with the Z-Image Turbo image generation model to enable Text-to-3D functionality. There is no need for input assets, making a small leap forward for ideation. Optionally, it also includes default support for Image-to-3D inference using direct image assets. Find the demo and related collections below... ๐Ÿค—๐Ÿ”ฅ

โœจ TRELLIS.2-Text-to-3D [Demo]: prithivMLmods/TRELLIS.2-Text-to-3D
โœจ Multimodal Collection: https://huggingface.co/collections/prithivMLmods/multimodal-implementations
โœจ Github: https://github.com/PRITHIVSAKTHIUR/TRELLIS.2-Text-to-3D

To know more about it, visit the app page or the respective model page!
prithivMLmodsย 
posted an update about 2 months ago
view post
Post
2049
Demo for Molmo2 on Hugging Face is live now, including Single/Multi-Image VQA, Visual Pointing/Grounding, Video VQA, and Video Point Tracking. Find the demo and related collections below. ๐Ÿ”ฅ๐Ÿค—

โ— Molmo2 HF Demo๐Ÿ–ฅ๏ธ: prithivMLmods/Molmo2-HF-Demo
โ— Model Collection: https://huggingface.co/collections/allenai/molmo2
โ— Related Multimodal Space Collection: https://huggingface.co/collections/prithivMLmods/multimodal-implementations

To know more about it, visit the app page or the respective model page!