Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
InstantStyle
Free Lunch towards Style-Preserving in Text-to-Image Generation
Tuning-free diffusion-based models have demonstrated significant potential in the realm of image personalization and customization. However, despite this notable progress, current models continue to
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
InstantID: Zero-shot Identity-Preserving Generation in Seconds
paper page:
model supports identity-preserving generation in high fidelity with only single reference image in any style
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
InstantID works with ControlNet Pose and LCM, and it might actually work with any ControlNet. The trade-off of using multiple ControlNets results in a slight loss of facial detail.
Now, you have SDXL-lightning on InstantID! For this example (faded film style), I think LCM-LoRA suffer less from style degradation (not a rigorous comparison).
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
OpenDiT is a great work by
@oahzxl
@zzk_zhao
@lzm_mlsys
from
@NUSingapore
, which is an Easy, Fast and Memory-Efficent System for DiT Training and Inference. This year will belong to DiT, you can't miss it if you are on generative boat.
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
I'm excited to share that our InstantID has been deployed on
@FEDML_AI
which provides the generative AI platform and foundation models to enable developers and enterprises to build and commercialize generative AI applications. Find more info at .
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
Image-based stylization has been supported in InstantStyle. Found more information at …. We will further combine InstantID for face stylization once stars reach 1K on GitHub.
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
A recent comparison about Distillation methods (4 steps with CFG=0). I have used these methods in my daily workflow, and I love SDXL-Lightning most for its good tradeoff between style degradation and image quality. Not sure whether TCD achieves the worst result in my test.
Now, you have SDXL-lightning on InstantID! For this example (faded film style), I think LCM-LoRA suffer less from style degradation (not a rigorous comparison).
Run InstantStyle Locally with 1 Click
InstantStyle lets you generate images with a style of ANY other image, instantly. No LoRA required. Both text-to-image/image-to-image.
I wrote a 1 click launcher for the gradio app from
@Haofan_Wang
(The author of InstantStyle/InstantId!).
We clarify that is quite misleading but not authorized and has never contacted us for official cooperation. Please pay attention to your personal privacy. We currently only have a project page, Github page, and a huggingface spaces demo.
#InstantID
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
We've gotten a lot of love❤️ from our users, and now you can support InstantID by buying us a cup of coffee via GitHub Sponsor (). More interesting projects are on the way.
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
Analyzing the contribution of each atomic layer is incomplete because different layers may influence each other. This is quite obvious in SD1.5, and only a few block mixtures will show obvious semantics. But like SDXL, these representation layers are all located near mid_block.
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
Introducing Face-to-All👨🎤, a diffusers 🧨 workflow inspired by
@fofrAI
amazing Face-to-Many ComfyUI workflow
Input a face, any style LoRA and get a stylized portrait
Colab with code:
Thanks
@Haofan_Wang
for merging our img2img pipeline to InstantID!
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
We are co-organizing a Spring Festival event with
@huggingface
from 2.7 to 2.25 on
@xiaohongshu
. Post your image with Spring Festival costumes, and win our official gifts. Happy Chinese 🐲 New Year! Happy Lunar New Year!
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
It's sad to know all plugins (LoRAs, ControlNets, Adapters, InstantID, etc) for playground-2.5 model need to be re-trained although it's also SDXL structure. A good news is that I have made a PR to well support playground-2.5 in diffusers.
🔥InstantID demo is now out on Spaces.
Thanks
@Haofan_Wang
et al, for building a brilliant Gradio demo for the community🙌
Check out the path-breaking demo now! Here is an example of a Marvel superhero, 🦸♂️
@ylecun
, generated using InstatID within seconds!
Renting an A100 (80GB) costs about $1,000 per month, which means you'd have to have 100 subscribers willing to pay $10 per month to break even. This is just the cost of GPUs.
Try this amazing tool () for generating professional headshots on the fly. To be honest, I don't want to upload dozens of my personal images to a website that uses my images for training, expensive, time-costing and unsafe.
🚀CharacterFactory: Sampling Consistent Characters with GANs for Diffusion Models
We propose CharacterFactory, a framework that allows sampling new characters with consistent identities in the latent space of GANs for diffusion models.
🔥Project page:
Thanks
@SiliconFlowAI
for their OneDiff integration of our InstantID! You can enjoy accelerated inference for InstantID (1.8x acceleration on RTX 4090). You can find more details at
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
Thanks
@MikeShou1
for hosting me at ShowLab, thanks
@YangYou1991
for arranging the guest lecture, and finally thanks the team from
@HPCAITech
for the hospitality tonight. Now, It’s time to take a look at my experiments and OpenAI’s Sora.
Coding tonight, lots of ideas to try. InstantID is not the end, not even the beginning of the end, our group are working very hard on other interesting projects, more details will be released gradually on . Run, don't walk, and always get prepared.
Do we need a stronger style model under the circumstance that we already have IP-Adapter, StyleDrop, StyleAlign and etc. Style reference on Midjourney is also good. But we cannot define a style accurately, right?
AnyV2V can bring Any Image Editing method to the video domain at no cost!
Now we have InstantStyle + AnyV2V!
AnyV2V
InstantStyle
Thanks
@Haofan_Wang
@vinesmsuic
for supporting!
Well, InstantID can already achieve this feature. The over-saturated result of InstantID reported in this paper is not fair, we are much better actually. Anyway, time will tell.
🤔Want to experience the power of InstantStyle?
🚀Code has been released, we'd welcome community contributed gradio demos!
💡Get inspired and start building your own style-preserving image generation apps with Gradio! Start here-
Decoupling content and style is promising. I trained B-LoRA locally on several samples using official training setting but cannot get satisfied results (middle columns). Not sure is there something I have missed.
Excited to share our new work B-LoRA🚀.
With our method, you can use a simplified version of LoRA trained on SDXL to perform style transfer between images and manipulate styles based on text.
Check out our website:
After testing,InstantID>IpadapterFaceID v2 >Photomaker in datasets testing
InstantID traiginng a controlnetxl model for position
i made windows, 16G VRAM need
Auto install and Download Model. it used SDXL model
Feture:
·install windows
·load local model
·support mac mps
My notes on Stable Diffusion 3 based on its generated results. It will become a new baseline soon.
1) Better text understanding capability with a new text encoder, maybe T5-XXL model. (1/n)
Announcing Stable Diffusion 3, our most capable text-to-image model, utilizing a diffusion transformer architecture for greatly improved performance in multi-subject prompts, image quality, and spelling abilities.
Today, we are opening the waitlist for early preview. This phase…
Zero-shot face-adapted image generation is a rapidly developing niche research field.
If you're looking to stay ahead of the curve or to simply exploring current possibilities with Gradio apps, this thread is the perfect place to start.
1⃣IPAdapter
2⃣PhotoMaker
3⃣InstantID
We support
#ControlNet
and
#T2IAdapter
both in
#diffusers
now!
A few comments for recent progress:
(1) The first released work has huge advantage over all other followers.
(2) The completness and usability are quite important.
I do love open source. But it sucks when I see some other companies adopting our work directly into their products for profit, ignoring the licenses you have.
Style is an underdetermined and mixed attribute that covers color, material, atmosphere, design, structure and so no. In some cases, it's even so tightly coupled to content that simply removing the content would break the style. But we can decouple part of cases at least.
I'm a big fan of IP-Adapter, which is an elegant but effective work. It has already performed pretty good on style consistency. But do you think we need a better style adapter than IP-Adapter?
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
If you have an advanced and practical model, go for closed-source product and make it profitable. If not, go for open-sourcing for popularity. The hardest thing is to find the critical point.
Thanks
@_akhaliq
for sharing our work!
TL;DR: We introduce InstantID as the state-of-the-art tuning-free method to achieve ID-Preserving generation with only single image.
Code:
Project Page:
The results look great. But it is not compatible with LCM-LoRA or SDXL-lightning in my local test. Any plan to support this model? We really need it in deployment.
@SimianLuo
1/ We are releasing Playground v2.5, our latest foundation model to create images.
We tested our model across 20K+ users in a rigorous benchmark that went beyond anything we've seen to date.
This model is open weights. More information in the tweets below. 👇
I plan to travel to Singapore with my family during the Chinese New Year. If you are a researcher, entrepreneur or investor and are interested in our recent work, drop me a message, I am happy to take a coffee and discuss.
Prerequisites for generative research
(1) A project page with fancy demos and potential applications.
(2) Fast integration into popular library such as diffusers, with support for common backbones such as SD1.5 and SDXL.
(3) Repost by AK on social media.
About two years ago, we tried to generate a coherent comic-like story by retrieving. We never knew it could be achieved with generative models and LLM then.
By
I'm excited to see that InstantID has been on Github and Paperwithcode's trending. At this moment, we are working very hard on diffusers integration and gradio demo, you will see both of them very very soon. So, don't forget to star our work.
Midjourney just released a new feature: Style Reference
It allows you to use images as references, similar to what the style tuner offered back in v5. However, the major difference between the two is that sref does not generate a "unique identifier" for your styles.
→ This…
Our work InstantStyle is lightweight and can be easily integrated into other tasks with a few lines, such as stylized video generation. Free feel to contact us for joint promotion on GitHub and Twitter👐
Thanks
@_akhaliq
for sharing!
TL;DR: InstantStyle is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
Code:
Project Page:
🔥InstantID demo is now out on Spaces.
Thanks
@Haofan_Wang
et al, for building a brilliant Gradio demo for the community🙌
Check out the path-breaking demo now! Here is an example of a Marvel superhero, 🦸♂️
@ylecun
, generated using InstatID within seconds!
Does the number of Github Stars really matter? As of now, PhotoMaker’s stars (6.1K) are greater than the sum of InstantID (1.8K) +IP-Adapter (2.8K). Can anyone tell me why?
Excited to share we just open-sourced our new background segmentation model 🥳
🚨 Check out our
@gradio
demo
RMBG v1.4 by BRIA excels in separating foreground from background across diverse categories, surpassing current open models 🚀
I visited Sanxindui museum this holiday and was really shocked by the splendid civilization at Bronze Age. Go back to work tomorrow, keep moving forward.
🖼 InstantID is now running on ✨ the non-profit GPU cluster 🥳
Thanks to
@QixunWang
❤ Xu Bai ❤
@Haofan_Wang
❤ Zekui Qin ❤ Anthony Chen ❤
🌐page:
🧿demo: please try it 🐣
We're introducing experimental support for `device_map` in Diffusers 🧪
If you have multiple GPUs you want to use to distribute the pipeline models, you can do so. Additionally, this becomes more useful when you have multiple low-VRAM GPUs.
Docs ⬇️
1/4
@airesearch12
@_akhaliq
Possibly, but we don't use any names in the prompt. To make the results more convincing, we also showed the results of ourselves as nobody.
InstantStyle demo is out! Upload the picture of an image and whatever you generate will come out in that style
You can choose style only blocks or style+layout! 🔥
We clarify that is quite misleading but not authorized and has never contacted us for official cooperation. Please pay attention to your personal privacy. We currently only have a project page, Github page, and a huggingface spaces demo.
#InstantID
Update about CVPR 2024:
1. One paper with 4(5), 4(3), 3(3). Hope to see you in Seattle if accepted.
2. One paper got desk reject, as our 1st author forgot to withdraw it from AAAI. But, 5 papers submitted to ICML this year, all thanks to my great interns.
@SimianLuo
I love your work, whether it gets accepted into any conference or not, bro. A similar case is IP-Adapter, they are both elegant and well-known works. Time will tell.