FontCLIP: A Semantic Typography Visual-Language Model for Multilingual Font Applications

paper review

paper-review
typography
vision-language-model
CLIP
multilingual
cross-lingual
semantic attributes
Published

Thursday, September 26, 2024

Keywords

manifold learning, font retrieval, font generation, typography, vision-language model, CLIP, multilingual, cross-lingual, semantic attributes

Abstract

Acquiring the desired font for various design tasks can be challenging and requires professional typographic knowledge. While previous font retrieval or generation works have alleviated some of these difficulties, they often lack support for multiple languages and semantic attributes beyond the training data domains. To solve this problem, we present FontCLIP: a model that connects the semantic understanding of a large vision-language model with typographical knowledge. We integrate typography-specific knowledge into the comprehensive vision-language knowledge of a pretrained CLIP model through a novel finetuning approach. We propose to use a compound descriptive prompt that encapsulates adaptively sampled attributes from a font attribute dataset focusing on Roman alphabet characters. FontCLIP’s semantic typographic latent space demonstrates two unprecedented generalization abilities. First, FontCLIP generalizes to different languages including Chinese, Japanese, and Korean (CJK), capturing the typographical features of fonts across different languages, even though it was only finetuned using fonts of Roman characters. Second, FontCLIP can recognize the semantic attributes that are not presented in the training data. FontCLIP’s dual-modality and generalization abilities enable multilingual and cross-lingual font retrieval and letter shape optimization, reducing the burden of obtaining desired fonts.

(Tatsukawa et al. 2024)

References

Tatsukawa, Yuki, I‐Chao Shen, Anran Qi, Yuki Koyama, Takeo Igarashi, and Ariel Shamir. 2024. “FontCLIP: A Semantic Typography Visual‐language Model for Multilingual Font Applications.” Computer Graphics Forum 43 (2). https://doi.org/10.1111/cgf.15043.