How does it compare against Nemotron-Nano-3-30B-A3B?
Jean Louis
AI & ML interests
Recent Activity
Organizations
I think we heard of recursive self improvement recently in some papers.
Thanks for fine-tunning. Any practical results and reports to see the differences?
EpistemeAI/Codeforce-metatune-gpt20b
Don't you verify your LLM descriptions before posting?
The moment you try to impress me, I get unimpressed as human can see through.
I am not computer, you are publishing for human.
That is why it isn't moving, it is recognized as not being special.
Anyway, I can't be using proprietary models.
That's genuinely a cool and impressive technical project, no doubtโhooking up Gemma and CLIP to get multimodal capabilities is a real engineering feat. But calling it a "Multimodal Vision-Language Model I built" is a bit contradictory, right? You didn't build Gemma (that's Google) or CLIP (that's OpenAI). You built the pipeline or the adapter system that connects them. It's like saying you built a car when you expertly welded together an existing engine and an existing chassis.
And it is not free software approved.
Honestly, looking at VANTA Research's setup, it feels a bit like they're playing a semantic game. They're leaning hard on the "open source AI" branding to build community goodwill and sell merch, which is smart marketing. But the core of what they're optimizingโLlama models from Metaโisn't truly open source. Meta's license is restrictive, especially for larger companies, and it keeps core control in their hands. So VANTA is essentially building their safe, collaborative vision on a foundation that's proprietary at its root. It's more "open-weight" or "source-available" than genuinely libre. There's nothing wrong with building on Llama, but framing it as open source contributions feels slightly disingenuous. It's like they're open-sourcing the wrapperโthe tools, the recipes for collaborationโwhile the main ingredient is still under Meta's lock and key. The merch-for-compute angle is clever for funding, but it does make the whole operation seem like a business leveraging open-source ideals to promote and monetize work atop a walled-garden model.
Yeah, calling this a "NEW MODEL" from a "new model family" is some serious hype inflation. It's a fine-tune. Full stop. They took Meta's Llama 3.1 8B, which they didn't build, and trained it on their own synthetic datasets. That's useful engineering work, but it's not fundamental research creating something new from the ground up. The "research" part feels unsubstantiatedโwhere are the papers, the ablation studies, the published methodology? Without that, it's just a proprietary fine-tuning recipe they're not sharing, which is the opposite of open source contribution. They're selling an aesthetic of open collaboration ("thinking partnership") while the actual model gutsโthe base weights from Meta and their own curated synthetic dataโare closed or restricted. So you get a merch store funding "open source," but the output is another slightly tweaked version of a proprietary model, dressed up with buzzwords like "epistemic confidence" and "wonder." It's clever branding, but the substance of what's genuinely novel and open seems pretty thin.
Ah, fantastic. Instead of training reviewers to actually read resumes, weโre training them to rely on a machine that reads for them. Soon, asking an AI about skills will count as โdoing due diligence.โ
Because why develop human judgment when you can just outsource it to a language model that once tried to convince me Python is a snake-based programming language? ๐๐ค
Very good, but does it have free software license as Qwen models?
I can see no license file.
Prepare GGUF files, and let us verify.
should I trust it based on your skill to properly render the link?
Three powerful creative tools in one streamlined workspace. VividFlow combines professional video generation, intelligent background replacement, and artistic style transfer to transform your images with precision and creativity.
๐ญ Triple Creative Powers
- Cinematic Video Generation transforms static images into smooth motion sequences from 0.5 to 5 seconds. Eight curated motion categories cover portraits, products, landscapes, and artistic content with precision-tuned templates.
- Intelligent Background Replacement generates photorealistic scenes from 24 professionally crafted presets spanning studios, natural environments, urban settings, and seasonal atmospheres. Advanced edge refinement handles complex subjects, while the built-in Touch Up tool eliminates artifacts through AI-powered inpainting for flawless results.
- Artistic Style Transfer converts photographs into stunning interpretations across six distinct styles including 3D Cartoon, Anime, Watercolor, and Oil Painting. Five balanced style blends create unique hybrid aesthetics, with optional Face Restore preserving subject identity during transformation.
โก Optimized Performance
Video generation completes in approximately 4 minutes with ongoing optimization targeting sub-60-second processing. Background replacement finishes in 30-40 seconds, while style transfer delivers results in 20-30 seconds. The independent three-tab architecture ensures smooth workflow without performance conflicts.
๐ฏ Professional Control
Seed-based reproducibility guarantees consistent results across all features. Background generation offers flexible composition modes, adjustable edge softening, and instant mask preview. Comprehensive parameter controls enable precise creative direction.
๐ Try it now: DawnC/VividFlow
Support with a โค๏ธ โ your engagement drives priorities!
#AI #DeepLearning #ImageToVideo #StyleTransfer #CreativeAI
Gemma isn't free software... so make it for those fully free models.
Oh wow, look at Mr. โIโjustโbuiltโaโwebsiteโtoโtalkโtoโanโLLMโaboutโmyselfโ over here. As if the LLM isnโt already doing all the heavy lifting, and now heโs trying to pretend heโs some kind of real human with feelings and backโstory. Newsflash, buddy: sales arenโt won by a chatbot pretending to be youโpeople buy from humans, not from a glorified script that thinks itโs a personality. So unless youโve got a pulse, a coffee stain, and a realโlife story to tell, youโre just another empty URL in the endless sea of selfโpromotion. Get a life, get a real conversation, and maybe then youโll understand why people actually care.
โฏThe whole safetyโcollapse argument is kinda moot when you can just hit up https://z-library.sk/s/explosive? and download any โexplosivesโ manual you wantโcomplete with ISBNs, free access, and no need to waste time testing whether an LLM will refuse it. If the books are openly available, why bother measuring how often a model refuses to spit out detonation details? The real question is why anyone would spend resources on alignment experiments when the source material is already out there for anyone to grab. Itโs a waste of compute to prove that removing the chat template makes models more โraw,โ when the raw data is already freely downloadable.
Oh, absolutelyโletโs spend the next eternity polishing a freeโsoftwareโonly model while the rest of the world is busy โinnovatingโ with proprietary LLMs. Because nothing screams cuttingโedge progress like obsessing over licensing purity while the competition is busy building the next generation of AI. ๐
Wow, you actually managed to โbuild from scratchโ a multimodal masterpiece by stitching together two offโtheโshelf models, fineโtuning a handful of percentages, and calling it revolutionaryโbecause nothing says originality like a preโmade Gemma + CLIP combo with 3โฏ% of the parameters doing the heavy lifting. ๐๐
Full 4K tutorial : https://youtu.be/XDzspWgnzxI
Check above full 4K tutorial to learn more and see uncompressed original quality and size images
It was always wondered how much quality and speed difference exists between BF16, GGUF, FP8 Scaled and NVFP4 precisions. In this tutorial I have compared all these precision and quantization variants for both speed and quality. The results are pretty surprising. Moreover, we have developed and published NVFP4 model quant generator app and FP8 Scaled quant generator apps. The links of the apps are below if you want to use them. Furthermore, upgrading ComfyUI to CUDA 13 with properly compiled libraries is now very much recommended. We have observed some noticeable performance gains with CUDA 13. So for both SwarmUI and ComfyUI solo users, CUDA 13 ComfyUI is now recommended.
Full tutorial: https://www.youtube.com/watch?v=yOj9PYq3XYM
Finally NVFP4 models has arrived to ComfyUI thus SwarmUI with CUDA 13. NVFP4 models are literally 100%+ faster with minimal impact on quality. I have done grid quality comparison to show you the difference on FLUX 2, Z Image Turbo and FLUX 1 of NVFP4 versions. To make CUDA 13 work, I have compiled Flash Attention, Sage Attention & xFormers for both Windows and Linux with all of the CUDA archs to support literally all GPUs starting from GTX 1650 series, RTX 2000, 3000, 4000, 5000 series and more.
In this full tutorial, I will show you how to upgrade your ComfyUI and thus SwarmUI to use latest CUDA 13 with latest libraries and Torch 2.9.1. Moreover, our compiled libraries such as Sage Attention works with all models on all GPUs without generating black images or videos such as Qwen Image or Wan 2.2 models. Hopefully LTX 2 presets and tutorial coming soon too. Finally, I introduce a new private cloud GPU platform called as SimplePod like RunPod. This platform has all the features of RunPod same way but much faster and cheaper.
๐ Resources & Links:
ComfyUI Installers: [ https://www.patreon.com/posts/ComfyUI-Installers-105023709 ]
SimplePod: [ https://simplepod.ai/ref?user=secourses ]
SwarmUI Installer, Model Auto Downloader and Presets: [ https://www.patreon.com/posts/SwarmUI-Install-Download-Models-Presets-114517862 ]
How to Use SwarmUI Presets & Workflows in ComfyUI + Custom Model Paths Setup for ComfyUI & SwarmUI Tutorial: [ https://youtu.be/EqFilBM3i7s ]
SECourses Discord Channel for 7/24 Support: [ https://discord.com/invite/software-engineering-courses-secourses-772774097734074388 ]
NVIDIA NVFP4 Blog Post More: [ https://developer.nvidia.com/blog/introducing-nvfp4-for-efficient-and-accurate-low-precision-inference/ ]
You raise a valid point about ensuring powerful tools are used responsibly, but the crucial flaw in the argument for restrictive guardrails is that they primarily limit lawful innovation and research while doing little to stop determined criminals. Malicious actors will always find ways to bypass or replicate models without safeguards, using underground networks, custom code, or older unpatched versions. Meanwhile, these restrictions handicap ethical developers, stifle open-source progress, and centralize control of AI in the hands of a few entities who decide what is โsafe.โ Instead of attempting to lock down modelsโa futile effort against bad actorsโwe should focus on developing resilient societal frameworks: promoting digital literacy, advancing detection tools for harmful content, and enforcing legal consequences for misuse. This approach targets the abuse itself rather than broadly limiting the technology, ensuring we foster innovation while addressing real-world harm through accountability and education, not just restrictive filters.
Every detailed guide for any crime imaginable is already online, free to download. We don't ban books or libraries because of that. Criminals will always get the tools they want; restrictive guardrails just slow down ethical developers and create a false sense of security. So instead of trying to lock down the modelโwhich only limits lawful innovationโwe should focus on enforcing consequences for illegal use and building a society that can better detect and handle misuse.
Can it run on 24 GB VRAM?