AI & ML interests

None defined yet.

Recent Activity

ginipickΒ 
posted an update 11 days ago
view post
Post
1958
🍌 Huggwarts Banana - Nanobanana + Long-form Text Rendering Add-on

🎯 The Critical Flaw in Nanobanana
Google's Nanobanana is excellent, but has one critical weakness:
Long-form text rendering breaks easily.
Huggwarts Banana = Nanobanana + Long-form Text Rendering Add-on

ginigen/AI

✨ Core Value
πŸ”₯ Nanobanana's Foundation

Text-to-Image generation
Multiple image models
Style customization
High-quality outputs

🎨 + Our Add-on Enhancement
100% Backward Compatible - Use it exactly like Nanobanana!

πŸš€ Extended Features
πŸ“Œ 1. Perfect Long-form Text Rendering

Natural line-breaking for lengthy sentences
Full multilingual support (Korean, English, Japanese, Chinese, etc.)
Complex character combinations (double consonants, compound vowels, special symbols) handled flawlessly
Mixed emoji 😊 and special characters β™₯
8 optimized fonts

🎬 2. Auto Image-to-Video Conversion
One-click video generation
Up to 12 seconds video length
Auto-applied effects: fade-in, zoom, slide transitions
SNS-optimized formats: 9:16 (Reels), 1:1 (Feed), 16:9 (YouTube)

🎨 3. Advanced Text Effects
Gradient colors: rainbow to subtle color transitions
Shadow effects: 3D depth perception
Neon glow: eye-catching luminous text
Custom backgrounds: AI-generated or custom upload

πŸ“ 4. Multiple Aspect Ratios
16:9 / 9:16 / 1:1 / 4:3, 3:4, 3:2, 2:3

🎯 5. Smart Model Selection
Automatic optimal model selection
Text-focused vs. Image-focused modes
Reference image support
High-resolution output (up to 4096x4096)


πŸ’‘ Simple 4-Step Workflow
1️⃣ Input Text β†’ English/Korean/Multilingual/Mixed content
2️⃣ Select Style β†’ Font, color, effects
3️⃣ Generate Image β†’ 10-30 seconds
4️⃣ Convert to Video (Optional) β†’ One-click automation

🎯 Use Cases
βœ… SNS Creators - Long-form text Reels/Shorts production
βœ… Marketers - Rapid text-based promotional images/videos
βœ… Educators - Visual text learning materials
NymboΒ 
posted an update 11 days ago
view post
Post
1504
Two new tools added to the Nymbo/Tools MCP server, File_System and Shell_Exec. You can theoretically do basically anything with these two tools, and it should enable support for many Claude Skills.

GPT-5-Codex proves that for many cases, shell commands really are all you need, and Claude Skills seem to lean into this. The thing is, nothing about the design of Claude Skills actually restricts them to proprietary models!

# File_System

There's a new directory inside the repo called Filesystem, that's the agent's "root". It can perform the following actions : list, read, write, append, mkdir, move, copy, delete, info, help. It's able to keep this all within the scope of one tool call by making the Action field required and all other fields optional. Using a filesystem shouldn't require 15 different tools.

Files created in the public HF space live in the space's running container, and gets cleared when the space is restarted. When running the server locally, files are actually stored on disk.

# Shell_Exec

What good is a filesystem if you can't execute commands in that filesystem? This tool automatically detects if the server is running on Windows or Linux, and suggests using the appropriate shell (PowerShell/Bash). Both of these new tools require that the agent uses relative paths, rather than absolute paths. I could be convinced to back pedal on this.

# Closing Thoughts

The File_System and Shell_Exec tools aren't super polished yet, I'll continue to improve the agent's instructions and UX of using the new tools. Most of my testing was done with gpt-oss-20b and if it messes up, it gets the gist after one failed tool call. It should work perfectly fine for the GPU poor.
  • 1 reply
Β·
multimodalartΒ 
posted an update 14 days ago
view post
Post
1767
Want to iterate on a Hugging Face Space with an LLM?

Now you can easily convert any HF entire repo (Model, Dataset or Space) to a text file and feed it to a language model!

multimodalart/repo2txt
NymboΒ 
posted an update 16 days ago
view post
Post
1693
I've made some improvements to my custom Deep_Research tool in the Nymbo/Tools MCP server. I've added a second LLM process and it still takes less than 1 minute to complete!

The original version of my Deep_Research tool would basically dump up to 50 fetched webpages onto the Researcher model (Qwen3-235B), with only a little bit of context shown from each page.

# New "Filterer" Process

The new process includes another LLM call before the researcher process. The Filterer (also Qwen3-235B) gets the query summary and the original 50 pages with low context, and decides which pages are most relevant to the research topic. The Filterer then outputs the URLs to the relevant pages, which are then re-fetched (with more context) and sent to the Researcher.

# Researcher Context

The Researcher now gets only the relevant webpages, then begins writing the report. When testing with 50 initial results, the researcher would often end up with 10-20 results of relevant context.

It still takes less than a minute to accomplish everything, thanks entirely to Cerebras inference. It now takes about 35-45 seconds to complete once the tool is run.

It's also worth noting that both the Filterer and Researcher now are provided the current time/date before they see the content, reducing hallucinations caused by knowledge cutoffs.
NymboΒ 
posted an update 26 days ago
view post
Post
615
I have a few Sora-2 invites - 15509N
  • 1 reply
Β·
NymboΒ 
posted an update about 1 month ago
view post
Post
1049
There's now a custom Deep_Research tool in my Nymbo/Tools MCP server! TL;DR: The agent using the tools writes a summary of your requests and up to five DuckDuckGo searches (up to 50 results). Each of the webpages found in the searches are then fetched and given to our researcher (Qwen3-235B-A22B-Thinking-2507). The researcher sees the summary, searched queries, and fetched links, then writes a thorough research report. The agent using the tool provides the user with a summary of the report and a link to download research_report.txt. The researcher's instructions are similar to some leaked Perplexity sys prompts.

# Deep_Research Tool

It accomplishes everything in under a minute so it doesn't hit MCP's 60 second timeout, mostly thanks to Cerebras. The only thing required to make this work is a HF_READ_TOKEN for inference.

The Deep_Research tool could certainly be improved. It still needs some sort of mechanism for sorting URLs based on importance (I've got some ideas but I don't want it to be the responsibility of the agent using the tool). I'll probably add a second researcher to filter out the bad sources before inferencing the big researcher. I'm hellbent on keeping this all within the scope of one tool call.

# More Fetch/Web Search Improvements

The Search_DuckDuckGo tool has been further enhanced. It now allows the agent to browse through all pages of results. The results also now include published date (if detected). It also now supports every DDG search types! Default DDG search is called text, but it can also now search by news, images, videos, and books.

The Fetch_Webpage tool now specifies how much of the page has been truncated, and cursor index, allowing it to pickup where it left off without re-consuming tokens. The model can now also choose to strip CSS selectors to remove excess noise, and there's a new URL Scraper mode that only returns URLs found on the full page.

More to come soon ~
ehristoforuΒ 
posted an update about 2 months ago
view post
Post
2137
πŸš€Hello from the Project Fluently team!

✨ We are happy to share with you our new universal LLM models based on Qwen3 1.7B and 4B β€” powerful, multilingual and ready to solve a wide range of problems!

πŸ› οΈ We have conducted additional training and carefully merged them to achieve even better results and maximize the potential of the models.

πŸ†“ And most importantly β€” the models are completely open and free under the Apache-2.0 license!

πŸ”— Links to repositories:
- FluentlyQwen3-4B: fluently/FluentlyQwen3-4B
- FluentlyQwen3-1.7B: fluently/FluentlyQwen3-1.7B

😍 We will be very glad to hear your feedback and impressions! Your opinion is very important to us!
NymboΒ 
posted an update about 2 months ago
view post
Post
1016
I have a few updates to my MCP server I wanna share: New Memory tool, improvements to web search & speech generation.

# Memory_Manager Tool

We now have a Memory_Manager tool. Ask ChatGPT to write all its memories verbatim, then tell gpt-oss-20b to save each one using the tool, then take them anywhere! It stores memories in a memories.json file in the repo, no external database required.

The Memory_Manager tool is currently hidden from the HF space because it's intended for local use. It's enabled by providing a HF_READ_TOKEN in the env secrets, although it doesn't actually use the key for anything. There's probably a cleaner way of ensuring memory is only used locally, I'll come back to this.

# Fetch & Websearch

The Fetch_Webpage tool has been simplified a lot. It now converts the page to Markdown and returns the page with three length settings (Brief, Standard, Full). This is a lot more reliable than the old custom extraction method.

The Search_DuckDuckGo tool has a few small improvements. The input is easier for small models to get right, and the output is more readable.

# Speech Generation

I've added the remaining voices for Kokoro-82M, it now supports all 54 voices with all accents/languages.

I also removed the 30 second cap by making sure it computes all chunks in sequence, not just the first. I've tested it on outputs that are ~10 minutes long. Do note that when used as an MCP server, the tool will timeout after 1 minute, nothing I can do about that for right now.

# Other Thoughts

Lots of MCP use cases involve manipulating media (image editing, ASR, etc.). I've avoided adding tools like this so far for two reasons:

1. Most of these solutions would require assigning it a ZeroGPU slot.
2. The current process of uploading files like images to a Gradio space is still a bit rough. It's doable but requires additional tools.

Both of these points make it a bit painful for local usage. I'm open to suggestions for other tools that rely on text.
ginipickΒ 
posted an update 2 months ago
view post
Post
4264
🍌 Nano Banana + Video: AI Image Style Transfer & Video Generation Tool

🎨 Key Features
1️⃣ Image Style Transfer

ginigen/Nano-Banana-Video

πŸ“Έ Upload up to 2 images for style fusion
✨ High-quality image generation with Google Nano Banana model
🎭 Apply desired styles with text prompts

2️⃣ Video Generation

🎬 Convert generated images to videos
πŸ“ Maintain original aspect ratio option
⏱️ Adjustable duration (1-4 seconds)

πŸš€ How to Use
Step-by-Step Guide
Step 1: Image Generation πŸ–ΌοΈ

Enter style description
Upload 1-2 images (optional)
Click "Generate Magic ✨"

Step 2: Video Creation πŸ“Ή

Send generated image to video tab
Set animation style
Generate video!

πŸ’‘ Use Cases

🏞️ Transform landscape photos into artistic masterpieces
πŸ€– Bring static images to life
🎨 Mix styles from two different images
πŸ“± Create short videos for social media

⚑ Tech Stack
Google Nano Banana Stable Video Diffusion Gradio Replicate API

#AIVideoGenerator #ImageToVideoConverter #StyleTransferAI #GoogleNanoBanana #StableVideoDiffusion #AIAnimationTool #TextToVideo #ImageAnimationSoftware #AIArtGenerator #VideoCreationTool #MachineLearningVideo #DeepLearningAnimation #HuggingFaceSpaces #ReplicateAPI #GradioApplication #ZeroGPUComputing #AIStyleMixing #AutomatedVideoProduction #NeuralStyleTransfer #AIPoweredCreativity
ginipickΒ 
posted an update 2 months ago
view post
Post
3661
πŸŽ‰ Fashion Fit 360: The New Standard in AI Virtual Try-On!

πŸš€ Now Live and Free to Use!Say goodbye to online shopping uncertainty - "Will this look good on me?" - with our revolutionary solution!Fashion Fit 360 is a cutting-edge AI-powered virtual fitting service that transforms your fashion shopping experience.

LINK: ginigen/Fashion-Fit360

✨ Core Features
πŸ”„ 360-Degree Multi-Pose Generation
Transform a single front-facing photo into 6 different viewing angles!
Front, side, and back views for complete visualization
Experience a real fitting room mirror effect
Check fit and style from every perspective

πŸ‘— 15 Fashion Item Categories
Apparel: Tops, bottoms, dresses
Jewelry: Necklaces, earrings, rings, bracelets
Accessories: Sunglasses, eyewear, hats, ties, bow ties, belts
Essentials: Bags, shoes

🎯 Perfect For:
πŸ›οΈ Online Shopping Enthusiasts: Preview before purchase - zero return hassles!
πŸ’ Jewelry Lovers: Virtually try expensive pieces before investing
🎁 Thoughtful Gift-Givers: Test items on recipient photos beforehand
πŸ‘” Business Professionals: Preview suit and tie combinations
πŸ‘— Fashion Designers: Rapidly visualize design samples

πŸ’‘ Why Fashion Fit 360?Fashion Fit 360 delivers innovation beyond conventional services.While most virtual fitting platforms only support clothing, we offer complete support for 15 accessory types. Unlike competitors providing only front views, Fashion Fit 360 generates 6 poses for true 360-degree visualization, ensuring you can verify actual fit perfectly.Performance is unmatched - get results in under 20 seconds with one-click simplicity and no complex configurations. Plus, download all generated images as a convenient ZIP file, eliminating tedious individual saves.

πŸ”₯ Key Differentiators
🎨 360-Degree Multi-Pose Image Generation
πŸ€– FLUX.1-Fill based OmniTry integrated model with Flux.1 KONTEXT LoRA technology
NymboΒ 
posted an update 2 months ago
view post
Post
1004
I built a general use MCP space ~ Fetch webpages, DuckDuckGo search, Python code execution, Kokoro TTS, Image Gen, Video Gen.

# Tools

1. Fetch webpage
2. Web search via DuckDuckGo (very concise, low excess context)
3. Python code executor
4. Kokoro-82M speech generation
5. Image Generation (use any model from HF Inference Providers)
6. Video Generation (use any model from HF Inference Providers)

The first four tools can be used without any API keys whatsoever. DDG search is free and the code execution and speech gen is done on CPU. Having a HF_READ_TOKEN in the env variables will show all tools. If there isn't a key present, The Image/Video Gen tools are hidden.

Nymbo/Tools
  • 1 reply
Β·
ginipickΒ 
posted an update 2 months ago
view post
Post
3402
✨ HairPick | Preview Your Perfect Hair Transformation in 360° ✨

🎊 Free Trial for Hugging Face Launch! Hurry! ⏰
Hello! Introducing an innovative AI service that helps you choose the perfect hairstyle without any regrets before visiting the salon!

🎯 Try It Now
ginigen/Hair-Pick

πŸ”„ What Makes HairPick Special? 360Β° Complete Preview!
Other hair simulators only show the front view? πŸ˜‘

HairPick is different!
βœ… Front + 4 random angles = Total 5 multi-angle images generated
βœ… Perfect check from side profile πŸ‘€ diagonal πŸ“ back view πŸ‘₯!
βœ… 100+ trendy hairstyle library πŸ’‡β€β™€οΈ

πŸ’‘ Highly Recommended For:
🎯 "I really don't want to fail this time!"
β†’ Check side volume and back lines thoroughly
🎯 "It's hard to explain exactly to my stylist"
β†’ Perfect communication with 360Β° result images!
🎯 "I have a profile photo/photoshoot coming up"
β†’ Preview your best look from every angle
πŸš€ Super Simple Usage (Just 1 Minute!)

1️⃣ One Selfie πŸ“Έ
Take a front-facing photo in bright light (show your forehead and face outline clearly!)
2️⃣ Choose Your Style πŸ’«
Select from 100+ options: short cuts, medium, long hair, layered, bangs, and more
3️⃣ Check 360Β° Results πŸ”„
Compare front + side + back + diagonal angles all at once!
4️⃣ Go to the Salon! βœ‚οΈ
Save your favorite result β†’ Show it to your stylist

πŸ“Έ Pro Tips for Perfect Results!
πŸ’‘ Lighting: Natural light or bright, even indoor lighting
πŸ’‘ Angle: Camera at eye level, facing straight ahead
πŸ’‘ Preparation: No hats❌ No sunglasses❌ Hair tucked behind earsβ­•

🎁 Now's Your Chance!
"The era of deciding based on front view only is over!"
HairPick isn't just simple hair synthesis, it's a next-level AI hair simulator that predicts your actual appearance in 360Β°.

πŸ”₯ Limited free access for Hugging Face launch!
πŸ”₯ 100+ latest trend styles!
πŸ”₯ ZERO failures with 360Β° perfect prediction!

βœ‚οΈ Click before you cut! Take on the perfect hair transformation with HairPick! 🌟

#HairPick #AIHairSimulator #360HairPreview
  • 2 replies
Β·
NymboΒ 
posted an update 2 months ago
view post
Post
1026
Anyone using Jan-v1-4B for local MCP-based web search, I highly recommend you try out Intelligent-Internet/II-Search-4B

Very impressed with this lil guy and it deserves more downloads. It's based on the original version of Qwen3-4B but find that it questions reality way less often. Jan-v1 seems to think that everything it sees is synthetic data and constantly gaslights me