Radamés Ajna

radames

AI & ML interests

None yet

Articles

Organizations

radames's activity

posted an update 4 days ago
view post
Post
1579
At Google I/O 2024, we're collaborating with the Google Visual Blocks team (https://visualblocks.withgoogle.com) to release custom Hugging Face nodes. Visual Blocks for ML is a browser-based tool that allows users to create machine learning pipelines using a visual interface. We're launching nodes with Transformers.js, running models on the browser, as well as server-side nodes running Transformers pipeline tasks and LLMs using our hosted inference. With @Xenova @JasonMayes

You can learn more about it here https://huggingface.co/blog/radames/hugging-face-google-visual-blocks

Source-code for the custom nodes:
https://github.com/huggingface/visual-blocks-custom-components
replied to their post 5 days ago
view reply

Yes that's a great idea, I'm chatting with folks from Convex, and check if the sqlite db is the only file I need to backup, then I'll set a scheduler to push it to a personal dataset. Folks them would be able to pause and restart from that state!

replied to their post 5 days ago
view reply

Thanks! It was a fun challenge to put it all together in a single container. I'm excited to try more Convexdb as a vector db and backend.

posted an update 6 days ago
replied to dhruvabansal's post 9 days ago
view reply

Just saying that I really like the ability to quickly test your model against the monster ones! It's amazing how well it performs against Claude. 🤯

replied to Xenova's post 11 days ago
view reply

Amazing!! Shall we make a VB node for this?

posted an update 11 days ago
view post
Post
2448
HiDiffusion SDXL now supports Image-to-Image, so I've created an "Enhance This" version using the latest ControlNet Line Art model called MistoLine. It's faster than DemoFusion

Demo: radames/Enhance-This-HiDiffusion-SDXL

Older version based on DemoFusion radames/Enhance-This-DemoFusion-SDXL

New Controlnet SDXL Controls Every Line TheMistoAI/MistoLine

HiDiffusion is compatible with diffusers and support many SD models - https://github.com/megvii-research/HiDiffusion
  • 1 reply
·
replied to Sentdex's post 18 days ago
replied to renyuxi's post 20 days ago
view reply

Hi @renyuxi , thanks for sharing this update! 8 steps with CFG and negative prompts is amazing!

posted an update 20 days ago
view post
Post
2415
I've built a custom component that integrates Rerun web viewer with Gradio, making it easier to share your demos as Gradio apps.

Basic snippet
# pip install gradio_rerun gradio
import gradio as gr
from gradio_rerun import Rerun

gr.Interface(
    inputs=gr.File(file_count="multiple", type="filepath"),
    outputs=Rerun(height=900),
    fn=lambda file_path: file_path,
).launch()

More details here radames/gradio_rerun
Source https://github.com/radames/gradio-rerun-viewer

Follow Rerun here https://huggingface.co/rerun
replied to oliveryanzuolu's post 27 days ago
posted an update 27 days ago
posted an update 27 days ago
replied to andrewrreed's post 28 days ago
view reply

Very interesting, @andrewrreed , and completely unaware of this feature! Do you know of any other strategies for grounded generation in models like LLaMA or Mistral?

posted an update about 1 month ago
posted an update about 2 months ago
view post
Post
2732
Following up on @vikhyatk 's Moondream2 update and @santiagomed 's implementation on Candle, I quickly put togheter the WASM module so that you could try running the ~1.5GB quantized model in the browser. Perhaps the next step is to rewrite it using https://github.com/huggingface/ratchet and run it even faster with WebGPU, @FL33TW00D-HF .

radames/Candle-Moondream-2

ps: I have a collection of all Candle WASM demos here radames/candle-wasm-examples-650898dee13ff96230ce3e1f
replied to freddyaboulton's post 2 months ago
view reply

nice!! can you set the jpeg quality as well?

replied to chansung's post 2 months ago
replied to Wauplin's post 2 months ago
posted an update 2 months ago
view post
Post
3498
Testing new pix2pix-Turbo in real-time, very interesting GAN architecture that leverages SD-Turbo model. Here I'm using edge2image LoRA single-step inference 🤯

It's very interesting how ControlNet Canny quality is comparable, but in a single step. Looking forward to when they release the code: https://github.com/GaParmar/img2img-turbo/issues/1

I've been keeping a list of fast diffusion model pipelines together with this real-time websocket app. Have a look if you want to test it locally, or check out the demo here on Spaces.

radames/real-time-pix2pix-turbo

Github app:
https://github.com/radames/Real-Time-Latent-Consistency-Model/

You can also check the authors img2img sketch model here

gparmar/img2img-turbo-sketch

Refs:
One-Step Image Translation with Text-to-Image Models (2403.12036)

cc @gparmar @junyanz
replied to visheratin's post 3 months ago
view reply

hi @visheratin , do you have any guides on how to train similar model? Phi-2 + SigLIP vision encoder?

replied to victor's post 4 months ago
view reply

very cool! I just ordered RPi5 to run some tests, also this awesome mic hat

image.png

replied to victor's post 4 months ago
view reply

I know it's possible to run real-time whisper on a rapberrypi with whisper.cpp @ggerganov

replied to victor's post 4 months ago
view reply

Are you thinking of running it on a device or in the cloud?

replied to abhishek's post 5 months ago