Ohad Fried
YOU?
Author Swipe
View article: Express4D: Expressive, Friendly, and Extensible 4D Facial Motion Generation Benchmark
Express4D: Expressive, Friendly, and Extensible 4D Facial Motion Generation Benchmark Open
Dynamic facial expression generation from natural language is a crucial task in Computer Graphics, with applications in Animation, Virtual Avatars, and Human-Computer Interaction. However, current generative models suffer from datasets tha…
View article: REED‐VAE: RE‐Encode Decode Training for Iterative Image Editing with Diffusion Models
REED‐VAE: RE‐Encode Decode Training for Iterative Image Editing with Diffusion Models Open
While latent diffusion models achieve impressive image editing results, their application to iterative editing of the same image is severely restricted. When trying to apply consecutive edit operations using current models, they accumulate…
View article: Differential Diffusion: Giving Each Pixel Its Strength
Differential Diffusion: Giving Each Pixel Its Strength Open
Diffusion models have revolutionized image generation and editing, producing state‐of‐the‐art results in conditioned and unconditioned image synthesis. While current techniques enable user control over the degree of change in an image edit…
View article: ImageRAG: Dynamic Image Retrieval for Reference-Guided Image Generation
ImageRAG: Dynamic Image Retrieval for Reference-Guided Image Generation Open
Diffusion models enable high-quality and diverse visual content synthesis. However, they struggle to generate rare or unseen concepts. To address this challenge, we explore the usage of Retrieval-Augmented Generation (RAG) with image gener…
View article: Tiled Diffusion
Tiled Diffusion Open
Image tiling -- the seamless connection of disparate images to create a coherent visual field -- is crucial for applications such as texture creation, video game asset development, and digital art. Traditionally, tiles have been constructe…
View article: Memories of Forgotten Concepts
Memories of Forgotten Concepts Open
Diffusion models dominate the space of text-to-image generation, yet they may produce undesirable outputs, including explicit content or private data. To mitigate this, concept ablation techniques have been explored to limit the generation…
View article: Stable Flow: Vital Layers for Training-Free Image Editing
Stable Flow: Vital Layers for Training-Free Image Editing Open
Diffusion models have revolutionized the field of content synthesis and editing. Recent models have replaced the traditional UNet architecture with the Diffusion Transformer (DiT), and employed flow-matching for improved training and sampl…
View article: The Chosen One: Consistent Characters in Text-to-Image Diffusion Models
The Chosen One: Consistent Characters in Text-to-Image Diffusion Models Open
Recent advances in text-to-image generation models have unlocked vast potential for visual creativity. However, the users that use these models struggle with the generation of consistent characters, a crucial aspect for numerous real-world…
View article: V-LASIK: Consistent Glasses-Removal from Videos Using Synthetic Data
V-LASIK: Consistent Glasses-Removal from Videos Using Synthetic Data Open
Diffusion-based generative models have recently shown remarkable image and video editing capabilities. However, local video editing, particularly removal of small attributes like glasses, remains a challenge. Existing methods either alter …
View article: Advancing Fine-Grained Classification by Structure and Subject Preserving Augmentation
Advancing Fine-Grained Classification by Structure and Subject Preserving Augmentation Open
Fine-grained visual classification (FGVC) involves classifying closely related sub-classes. This task is difficult due to the subtle differences between classes and the high intra-class variance. Moreover, FGVC datasets are typically small…
View article: Monkey See, Monkey Do: Harnessing Self-attention in Motion Diffusion for Zero-shot Motion Transfer
Monkey See, Monkey Do: Harnessing Self-attention in Motion Diffusion for Zero-shot Motion Transfer Open
Given the remarkable results of motion synthesis with diffusion models, a natural question arises: how can we effectively leverage these models for motion editing? Existing diffusion-based motion editing methods overlook the profound poten…
View article: DiffUHaul: A Training-Free Method for Object Dragging in Images
DiffUHaul: A Training-Free Method for Object Dragging in Images Open
Text-to-image diffusion models have proven effective for solving many image editing tasks. However, the seemingly straightforward task of seamlessly relocating objects within a scene remains surprisingly challenging. Existing methods addre…
View article: Break-A-Scene: Extracting Multiple Concepts from a Single Image
Break-A-Scene: Extracting Multiple Concepts from a Single Image Open
Text-to-image model personalization aims to introduce a user-provided concept to the model, allowing its synthesis in diverse contexts. However, current methods primarily focus on the case of learning a single concept from multiple images …
View article: Diffusing Colors: Image Colorization with Text Guided Diffusion
Diffusing Colors: Image Colorization with Text Guided Diffusion Open
The colorization of grayscale images is a complex and subjective task with significant challenges. Despite recent progress in employing large-scale datasets with deep neural networks, difficulties with controllability and visual quality pe…
View article: The Chosen One: Consistent Characters in Text-to-Image Diffusion Models
The Chosen One: Consistent Characters in Text-to-Image Diffusion Models Open
Recent advances in text-to-image generation models have unlocked vast potential for visual creativity. However, the users that use these models struggle with the generation of consistent characters, a crucial aspect for numerous real-world…
View article: Blended Latent Diffusion
Blended Latent Diffusion Open
The tremendous progress in neural image generation, coupled with the emergence of seemingly omnipotent vision-language models has finally enabled text-based interfaces for creating and editing images. Handling generic images requires a div…
View article: Differential Diffusion: Giving Each Pixel Its Strength
Differential Diffusion: Giving Each Pixel Its Strength Open
Diffusion models have revolutionized image generation and editing, producing state-of-the-art results in conditioned and unconditioned image synthesis. While current techniques enable user control over the degree of change in an image edit…
View article: Break-A-Scene: Extracting Multiple Concepts from a Single Image
Break-A-Scene: Extracting Multiple Concepts from a Single Image Open
Text-to-image model personalization aims to introduce a user-provided concept to the model, allowing its synthesis in diverse contexts. However, current methods primarily focus on the case of learning a single concept from multiple images …
View article: Deep Image Fingerprint: Towards Low Budget Synthetic Image Detection and Model Lineage Analysis
Deep Image Fingerprint: Towards Low Budget Synthetic Image Detection and Model Lineage Analysis Open
The generation of high-quality images has become widely accessible and is a rapidly evolving process. As a result, anyone can generate images that are indistinguishable from real ones. This leads to a wide range of applications, including …
View article: Prediction of Scene Plausibility
Prediction of Scene Plausibility Open
Understanding the 3D world from 2D images involves more than detection and segmentation of the objects within the scene. It also includes the interpretation of the structure and arrangement of the scene elements. Such understanding is ofte…
View article: FakeOut: Leveraging Out-of-domain Self-supervision for Multi-modal Video Deepfake Detection
FakeOut: Leveraging Out-of-domain Self-supervision for Multi-modal Video Deepfake Detection Open
Video synthesis methods rapidly improved in recent years, allowing easy creation of synthetic humans. This poses a problem, especially in the era of social media, as synthetic videos of speaking humans can be used to spread misinformation …
View article: Taming Normalizing Flows
Taming Normalizing Flows Open
We propose an algorithm for taming Normalizing Flow models - changing the probability that the model will produce a specific image or image category. We focus on Normalizing Flows because they can calculate the exact generation probability…
View article: Neural Font Rendering
Neural Font Rendering Open
Recent advances in deep learning techniques and applications have revolutionized artistic creation and manipulation in many domains (text, images, music); however, fonts have not yet been integrated with deep learning architectures in a ma…
View article: SpaText: Spatio-Textual Representation for Controllable Image Generation
SpaText: Spatio-Textual Representation for Controllable Image Generation Open
Recent text-to-image diffusion models are able to generate convincing results of unprecedented quality. However, it is nearly impossible to control the shapes of different regions/objects or their layout in a fine-grained fashion. Previous…
View article: Ham2Pose: Animating Sign Language Notation into Pose Sequences
Ham2Pose: Animating Sign Language Notation into Pose Sequences Open
Translating spoken languages into Sign languages is necessary for open communication between the hearing and hearing-impaired communities. To achieve this goal, we propose the first method for animating a text written in HamNoSys, a lexica…
View article: GEFF: Improving Any Clothes-Changing Person ReID Model using Gallery Enrichment with Face Features
GEFF: Improving Any Clothes-Changing Person ReID Model using Gallery Enrichment with Face Features Open
In the Clothes-Changing Re-Identification (CC-ReID) problem, given a query sample of a person, the goal is to determine the correct identity based on a labeled gallery in which the person appears in different clothes. Several models tackle…
View article: Blended Latent Diffusion
Blended Latent Diffusion Open
The tremendous progress in neural image generation, coupled with the emergence of seemingly omnipotent vision-language models has finally enabled text-based interfaces for creating and editing images. Handling generic images requires a div…