ids
stringlengths
36
36
texts
stringlengths
212
1.03k
4f2034a7-ac6c-489a-881d-3aef4f1d0c0d
Industrial Language-Image Dataset (ILID): Adapting Vision Foundation Models for Industrial Settings Keno Moenck1,*Duc Trung Thieu1Julian Koch1Thorsten Sch ¨uppstuhl1 1Hamburg University of Technology, Institute of Aircraft Production Technology github.com/kenomo/ilid In recent years, the upstream of Large Language Models (LLM) has also encouraged the computer vision community to work on substantial multimodal datasets and train mod- els on a scale in a self-/semi-supervised manner, resulting in Vision Foundation Models (VFM), as, e.g., Contrastive Language–Image Pre-training (CLIP). The models gener- alize well and perform outstandingly on everyday objects or scenes, even on downstream tasks, tasks the model has not been trained on, while the application in specialized do- mains, as in an industrial context, is still an open research question.
1db71561-d2b5-4705-9d1a-510b84df1992
Here, fine-tuning the models or transfer learning on domain-specific data is unavoidable when objecting to adequate performance. In this work, we, on the one hand, introduce a pipeline to generate the Industrial Language- Image Dataset (ILID) based on web-crawled data; on the other hand, we demonstrate effective self-supervised trans- fer learning and discussing downstream tasks after train- ing on the cheaply acquired ILID, which does not neces- sitate human labeling or intervention. With the proposed approach, we contribute by transferring approaches from state-of-the-art research around foundation models, trans- fer learning strategies, and applications to the industrial domain. 1. Introduction Machine vision technologies facilitated by deep learning usually outperform traditional methods, especially in dy- namic and open settings.
bf8e7d1a-62eb-43ac-a9b5-16e6796b80e8
In the scope of training deep mod- els, industrial contexts1lack everyday objects and scenes, typically covered by publicly available datasets, which is why applications in these specialized domains here demand *Corresponding author: keno.moenck@tuhh.de 1We define the industrial domain as follows: industrial activities serve to produce consumable goods or a capital asset, which includes production as the superordinate term involving all processes around it, including activ- ities from manufacturing, assembly, logistics, or finance. In addition, tasks in the later lifecycle of a product, like Maintenance, Repair, and Overhaul (MRO), also belong to industrial activities. Vision applications are typi- cally closer to the shopfloor than the topfloor.custom datasets, e.g., synthetically generated [1–4], which model the specific object and sensor domain.
bbdcb2e9-d7eb-4157-959f-38d1e2a4c0f5
The availability of curated, publicly accessible datasets specific to industrial needs is exceedingly sparse, e.g., the MVTec [5–8], VISION [9], or tool recognition [10] datasets encapsulate only a limited spectrum of objects and sup- port only a handful of trainable tasks based on the provided ground truth annotations. Besides the need for training data, fine-tuning, domain adaptation, or transfer learning, trans- ferring a model from a source to a related target, e.g., ob- ject/scene/sensor domain, is ineluctable, which can reduce the necessary samples per conceptual class to only a few shots during training. The model’s pre-training is the crit- ical point here, where training data size, variability, and model size directly relate to the overall performance [11]. Large-scale pre-trained foundation models represent a paradigm shift in Artificial Intelligence (AI), characterized by extensive self-supervised training [12].
2cd18c18-9a39-419d-99e6-a944263c0bb5
These models, e.g., BERT [13], the well-known GPT-n series [14–16], or Llama [17–19], learn rich knowledge representations capa- ble of transcending to various downstream tasks. The shift in AI drives single tasks and single-modalities learners to a paradigm encompassing diverse tasks and multimodalities, which more closely mimics human perception and cogni- tive processes. Following Large Language Models (LLM), Vision Foundation Models (VFM) have been upstreamed in the last few years, capable of supporting not only 2D or even 3D modalities but also language [20]. Data for training at scale is typically web-crawled from the vast re- sources of the Internet, which then demands sophisticated post-processing, posing a variety of challenges [21, 22].
25991450-a7f3-4f85-a8be-0e2832a4b0fd
Be- sides, given such large, partially unstructured datasets, only self-supervised or unsupervised methods are able to learn from the data effectively. A self-supervised approach capable of learning from text and image modalities is contrastive learning, in which a model learns to distinguish between positive and negative combinations of samples, firstly, nearly concurrently, pre- sented by CLIP [23] and ALIGN [24] at a large scale. Con- 1arXiv:2406.09637v1 [cs.CV] 14 Jun 2024
525e99cf-50bc-4f82-8020-5791a0cd2a22
trastive learning by contrasting positive and negative sam- ples in a batch, in the case of vision and language, is based on a text and image encoder. The idea is that the encoders are trained to output embeddings for the image and text, in- creasing the similarities of positive samples by decreasing the distance in the joint embedding space and increasing the distance of negative samples. Employing a text encoder allows for natural language supervision, relaxing the neces- sity of fixed classes as in the case of training traditional deep learning models like a ResNet [25]. This fact makes as- sembling a dataset at a scale less laborious since assigning an image to a fixed class omits, enabling learning from un- structured data. Different language-image datasets of scale have emerged, ranging from 12M [22] to 5B [21] samples.
29480e67-4a8c-45bd-8e09-382af24a9a92
Since they are based on web-available data, not all cleaned, post-processed, and curated datasets are published, as in the case of CLIP. „… levelling feet round “ 0.64 „… collet “ 0.24 „… aluminium profile “ 0.05 „… button “ 0.03 „…“ … „… button “ 0.33 „… collet “ 0.22 „… magnetic ball joint “ 0.22 „… axial joint “ 0.10 „…“ … (a) Ours (b) Zero - shot CLIP (baseline) Prompt Prompt Score Score Figure 1.
71f07a77-fb41-497a-ad90-e342aecd5cd6
CLIP on the task of classification after (a) transfer learn- ing on the Industrial Language-Image Dataset (ILID) and (b) the zero-shot baseline results. VFMs exhibit rich knowledge representations, are adapt- able to various downstream tasks, and generalize better than conventional models, but only to a certain extent in novel and out-of-distribution domains, necessitating fine-tuning or transfer learning. As demonstrated in Fig. 1, the zero- shot model CLIP, given a highly out-of-distribution image, does not predict nor even close to the ground truth. As already outlined, in the industrial domain, we face non- everyday objects and scenes, which is why we can not rely on commonly available datasets for fine-tuning or transfer learning, which also inhibits the use of VFM here.
e589ab17-f7e4-46b5-91da-6eeab7db4980
In this work, we try to make a step in the direction of utilizing VFM capabilities in specialized industrial domains by con- tributing three-folded: • We propose a method to generate the IndustrialLanguage-Image Dataset (ILID) from web-crawled data and release a version that covers objects from different industrial-related domains2. We publish the pipeline to generate the ILID at github.com/kenomo/ilid. • We effectively demonstrate transfer learning to CLIP with the given dataset, which outperforms CLIP’s zero-shot capabilities. • We elaborate on different tasks that serve indus- trial domain-related vision applications. We pub- lish the training- and evaluation-related code here github.com/kenomo/industrial-clip. This work focuses on utilizing CLIP rather than other vision-language models due to the significant established usage and fine-tuning/transfer learning strategies.
4c4b78fe-25b8-45cb-9b4a-86b79f805c54
Besides, comparing only one established model on the data increases the focus, clarity, and depth of the findings in the scope of this work. Nevertheless, we encourage the reuse of ILID with other strategies or also employ further fine-tuning and transfer learning strategies. The rest of this work is structured as follows: First, we outline in Sec. 2 existing applications of VFMs in industrial applications, introduce Contrastive Image-Language Pre- training (CLIP) and current existing fine-tuning/transfer- learning approaches. In Sec. 3, we present our overall method of generating the dataset as well as our training pro- cedure. We elaborate on our extensive experiments in Sec. 4. We conclude and discuss this work in Sec. 5. 2. Related Works 2.1.
b4a9809a-e533-4111-88e8-ef4bd77d466d
VFMs in industrial applications Code recognition, object or position recognition, complete- ness, shape/dimension check, or quantitive or qualitative inspection are typical vision applications in manufacturing [26]. While in manufacturing, these are often suited toward narrow fields of view and close to the object; in the neigh- boring domain, intralogistics, tasks are besides close ones, like inspecting load carriers for trash, contamination, and damage or documentation, verification, assistance, and au- tomation, perceiving the environment is often of interest, which results in, e.g., foreign debris detection or tracking objects [27, 28]. The first step in the perception pipeline of these applications is typically a fundamental vision task, e.g., in the 2D domain, giving each pixel semantic and clus- tering pixel to semantically meaningful regions. Then fol- lows additional enhancing the output with further semantics and finally forming the application-specific decision used in, e.g., part of a production system.
34dca95d-b4e3-4aa5-88da-46ff6d073f03
2Since the data from the web do not belong to us, we are not allowed to publish the images and texts, but we provide the final post-processed metadata, which can be used to reassemble the dataset. Please contact the corresponding author. 2
b9b7dfea-5a24-475a-bb5a-490955f095f8
(3) (Downstream) tasks (1) Industrial Language - Image Dataset (ILID) Web catalog crawling Dataset processing Image Image Classification Segmentation „…hinge…“ „…handle…“ „… rod end…“ „…“ Text Encoder Image Encoder maximizing the score for non - contrasting samples { bores for counter - sunk screws } { h inge, detachable } (2) Transfer learning Text [ { label_short : „hinge“, label_long : „hinge with countersunk bores“, description: „ Detachable hinge with high - quality die - casting process “, material: „ zinc die casting “, material_finish : „powder coated, black“, … }, { … }, … ] [ { label: „ Hinge M zinc die casting,
a806b244-2c38-443b-80f2-9697786a49e8
detachable , type R, xmm “ , data: [ „Hinges are distinguished by their compact and robust construction. The assortment of materials...”, „Zinc die casting ...”, „...“, ] }, { … }, … ] TextFigure 2. Overview of this work’s method: (1) generation of the Industrial Language-Image Dataset (ILID), (2) transfer learning using the ILID, and (3) evaluating the performance in different tasks. Up to this date, there exists only a small set of publica- tions on the topic of utilizing VFMs in one or more steps of such vision pipelines, which we give a small excerpt in the following: On a broader scale [29] explore use cases for deploying VFMs in the industrial context without de- signing or elaborating on specific architectures and how to train, fine-tune, or do transfer learning.
53a75ba1-4243-475f-9a64-ddb1fd3f6b2c
[28] discusses the abilities of the Segment Anything Model (SAM), a class- agnostic segmentation model that outstandingly generalizes to unseen objects and scenes, in the context of vision ap- plications in the aircraft industry, including manufacturing, intralogistics, and MRO. [30] name two use cases in PCB defect inspection and industrial human action recognition. Current literature throws up ideas on utilizing LLMs, e.g. [31], or VFMs, e.g., [28–30, 32], in the industrial do- main; little is known about how to enable VFM to perform effectively in specific use cases. Besides having suitable datasets, training with the data demands specific strategies. We will elaborate on the aspects in the following sections. 2.2. Contrastive Language-Image Pre-training (CLIP) CLIP learns rich image-text representations from natural language supervision utilizing natural language as a pre- diction space to reach higher performance in generalization and transfer.
b875aca6-6228-4e2b-b387-4d0bd9d96e3a
It is not an entirely novel approach; however,the origin of the idea of learning from perceptions in natural language is not exactly dated to specific research. In 1999, [33] explored retrieving words for unknown images based on statistical learning to predict nouns and adjectives. In 2007, [34] demonstrated learning image representations us- ing manifold learning to predict words for images. Recent approaches that emerged before CLIP and learn visual rep- resentations from text are Visual representations from Tex- tual annotations (VirTex) [35], Image-Conditioned Masked Language Modeling (ICMLM) [36], and Contrastive Visual Representation Learning from Text (ConVIRT) [37]. 2.2.1 Contrastive learning A contrastive learning model consists of two main compo- nents: (1) an encoder for all input modalities and (2) a loss function measuring the similarity between positive and neg- ative pairs.
aed90d94-616b-44e3-8402-7f73f488c913
The encoder can be reused from other models and training, e.g., demonstrated by OpenScene [38], which employs a frozen text and 2D image encoder while training a 3D point cloud encoder for language-based 2D/3D scene understanding. The encoder models are trained to comple- ment and comprehend each other fully by encoding similar concepts of images and text in similar vectors. That is, a text representing ”photo of a hinge” would output a similar vector as the image counterpart and be further away from 3
ad48f083-191b-4bb7-8906-d06fd1105b3a
images that are not connected, as shown in Fig. 3. Besides prompting for the object’s name, a sufficiently trained text encoder would encode, e.g., conceptual close activities near the object’s name embedding (s. Fig. 3). {photo of a hinge} {cat} {house} {gripping} Figure 3. Joint embedding space of text and image representations: conceptually similar texts and images are encoded close to each other, dissimilar pairings do not share similar positions. The self-supervised pre-training of CLIP followed: Given Npairs of image and text, CLIP estimates the sim- ilarity for all the possible N×Npairings. With each text and image pair in the multimodal vector space, the models inside CLIP are jointly trained to maximize the similarity of each positive pairing and, at the same time, minimize the similarity of N×N−Nantagonistic pairs (s. Fig. 2).
50dd7115-c548-4643-93c9-61ad337f8d0a
The embedding similarities between pairs are represented by the cosine similarity metric, which is used to optimize the cross-entropy loss in order to build the most optimized versions of both the image and text encoder at the same time. 2.2.2 Performance Zero-shot CLIP achieves similar performance or even out- performs conventional fully supervised class-wise models while preserving robustness through the ability to learn from a broader range of representations from natural lan- guage, especially on in-distribution or slightly out-of- distribution data. On the other hand, zero-shot CLIP weakly performs on datasets that are far out-of-distribution, such as satellite images (EuroSAT [39], NWPU-RESISC45 [40]) or tumors (PatchCamelyon [41]) [23].
834949f2-3bdf-42af-a550-565095b05870
When comparing CLIP and other large pre-trained n-shot models such as BiT-M [42] and SimCLRv2 [43], CLIP’s authors depict that the zero-shot performance outperforms all other models on the metric of average accuracy score up to 4-shot linear classi- fiers trained on the same dataset [23]. The limitations arethat scaling the model to learn from more data has steadily increased the performance, but computing power increases exponentially, which is currently barely economically rea- sonable. 2.2.3 Recent development Meanwhile, much work exists on further development and adaptations of CLIP [44–51]. The most notable works are SLIP [50], DeCLIP [44], ReCLIP [46], CoCa [47], and FILIP [49], aiming to improve efficiency in the train- ing process. SLIP combines language supervision and im- age self-supervision to improve performance further.
031e103c-a527-4ce7-9ff4-35108a47f1f4
De- CLIP employs supervision across modalities as well as self- supervision within each modality, whereas ReCLIP at first learns pseudo labels and then applies cross-modality self- supervision. CoCA, on the other hand, skips cross-attention in some decoder layers to encode unimodal text represen- tations and cross-attend the remaining layers with the im- age encoder. By using contrastive loss between unimodal image and text embeddings, along with captioning loss for multimodal decoder outputs, CoCa efficiently trains on a wider variety of data with minimal overhead. Improved fine-grained performance of CLIP is demonstrated in the works of FILIP [49], where instead of contrastive loss be- ing calculated from global features of an entire image and text sequence, token-wise cross-modal interaction is mod- eled to take into account image patches and textual tokens more fine-grained.
99376fdf-7135-4b57-a274-4e64cd677111
Since this work focuses mainly on the training data, we will not evaluate all the individual strategies that aim to in- crease performance. Instead, we use the vanilla CLIP model and employ basic transfer learning methods that we can em- ploy with limited hardware resources, which also demon- strate the effectiveness in the scope of lower-cost applica- tions. 2.3. Fine-tuning and transfer learning Depending on the application, CLIP has two different ways to adapt to a new distribution, i.e., new sets of data entirely outside the dataset on which CLIP was pre-trained. Fine- tuning and transfer learning are very similar ways to adapt CLIP, but they have different applications depending on the task at hand and different processes in modifying the archi- tecture. Fine-tuning consists of training all layers or at least parts of the model.
8322bfaf-71c2-4ec1-b80a-9a498ec4476c
This process is usually more suitable for adapting to small sets of data that are closely related to the dataset CLIP was pre-trained on, such as everyday objects and general concepts. On the other hand, in tasks where the dataset is too specific, i.e., specialized knowledge, transfer- learning is better suited, as it freezes all the original layers of the pre-trained model and only adds or injects extra train- able layers or parameters. This way, the learned features of 4
3925cbe7-e2c3-4e9b-a56c-d12057dfe45f
(2) Web crawling (3) Pre - filtering (4) Processing (5) Post - filtering (6) Downloading (1) Online catalogs [ {…}, {…}, … ] LLM [ {…}, {…}, … ] Figure 4. Dataset generation pipeline resulting in the Industrial Language-Image Dataset (ILID). the zero-shot model are preserved and optimized for gen- eralization to novel, previous out-of-distribution data. Usu- ally, the fine-tuning process requires much more resources in terms of time, data, and computation as it modifies all the layers of the model compared to transfer learning. In the case of ILID, transfer learning proved to be a fitting so- lution, as the dataset is specialized specifically on industrial components, which are not presumably contained in CLIP’s dataset used for pre-training.
f1c46544-8d6d-41fa-8fd4-7e9dd99463b7
Notable works in transfer-learning of CLIP are adapter- styled tuning, e.g., CLIPAdapter [52], and prompt learning, e.g., CoOp [53] and APEX [54]. CLIPAdapter (s. Fig. 5) adds dense down- and up-sampling layers on top of CLIP either to the image, text, or both encoders. Thereby, only the most prominent features are compressed into lower di- mensions. From the latent space, the adapter then learns to reconstruct the essential ones. CoOp (s. Fig. 5) is the first to demonstrate continuous prompt learning as in- troduced by [55] for CLIP, which is learning continuous prompts by backpropagation for each label or one specific prompt template for all labels.
475e70ad-9476-4c0f-9199-45f497467194
Concretely, CoOp creates a set of learnable vectors, initialized by random values or given text embeddings, which, during training, the model adapts to. APEX is the most recent approach that also eval- uates adding learnable tokens to the residual transformer blocks in the image encoder. Besides, APEX introduces a residual connection skipping the text adapter steered by an adaptive coefficient to perform better on a variety of out-of- distribution data. 3. Method In this section, we first outline the generation of the ILID, including a thorough outline of the dataset acquisition, the criteria for data selection, web crawling to gather extensive sets of unlabeled data, and filtering (s. Sec. 3.1). Secondly, we elaborate on the decision for the model architecture and training procedure in Sec. 3.2. 3.1.
a08a7bd6-0dc8-49b2-a50b-c45bcff0d968
Dataset generation pipeline Following a typical data pipeline structure, including data selection, transforming, and pre-/post-filtering (s., e.g.,[22]), we employed six steps (s. Fig. 4) to generate the In- dustrial Language-Image Dataset (ILID). Each of the steps results in a structured JSON document containing all the outputs. The next step always takes the respective docu- ment as input. 1. While searching for reasonably organized industrial- related data on the Internet, we found that online cata- logs contain relevant language-image information. Typ- ically, web stores have a page per product, sometimes imaging a set of product configurations, a precise, of- ten standardized, title, description, information about the material, and further information about the product. These online stores are an adequate data source for the industrial domain. The first step was identifying a store set containing the necessary object-domain.
fa851881-cb33-4fbe-8ec8-be2080599eeb
2.Web crawling data from online catalogs follows two ba- sic steps: getting the sitemap from robots.txt and writing a crawler for the specific structure of the product pages. The top-level robots.txt file delineates the Robots Exclu- sion Protocol , which guides crawlers and other bots on which sections of the website they are permitted to ac- cess. Typically, this file also specifies the location of the sitemap, an XML-formatted document designed to provide crawlers with information about all pages on a website. Sitemaps can be hierarchically ordered; in the case of online catalogs, typically, there is one specific sitemap containing all products and their respective lo- cations. We use Scrapy3as a Python-based web crawler that takes a sitemap as input and crawls through all the specified locations. Creating a specific spider for a web catalog requires manual intervention since one has to de- fine which images and text blocks to yield.
72e39c3d-be61-4221-bc66-a394bd766afd
Besides a central label tag for each entry, we save an unstructured list-typed data object, which can contain all other avail- able information about the product, like materials, finish, colors, etc. Using the sitemap as the initial crawling en- try point is a common step in every online search engine. 3. In the pre-filtering step, we filter for duplicate entries, remove special characters, as well as diminish entries that do not have sufficient information. Besides, we filter the data for a set of trade names and remove these from all product information. Often, industrial product names include the manufacturer, which we do not want to use further or bias the data within the following information extraction. 4. In the central processing step, we use a small local- deployable LLM to extract our five target information from the unstructured data.
0e657c86-7ced-4273-830b-b4c34fe59fa3
We define these as (1) a long label describing the product, (2) a short label that is shorter than the long label, (3) a description of the prod- uct, (4) the material, (5) the finish or color of the product (s. also Fig. 2). In our study, we used Llama3-8B [19] 3Scrapy: A Fast and Powerful Scraping and Web Crawling Framework 5
efea53e0-fdc6-4cda-8645-4442bffde53c
in the fine-tuned instruct version (s. 6 for the respective prompt). We ask the LLM not to output any numbers or sizes; additionally, we remove them from the initial data since, on the one hand, we do not expect that a 2D image task can identify or recognize any dimensional quantities given different camera positions and varying intrinsics, on the other hand, we do not want to bias the dataset with it. After prompting for the desired information, we extract these from the response and save them for further processing. We discard the item from the dataset if the prompt does not return sufficient output. 5. In the post-filtering step, we again filter for any un- wanted characters and do some further cleaning, like lowering words. 6. In the final downloading step, all images are down- loaded, post-processed, and resized while also assem- bling the final JSON specifying the dataset’s text and metadata.
4c4ab7e9-826f-4c0d-833c-a3e09d3c115a
With the given steps, we are able to extract a product’s image and a structured set of five pieces of information. Be- sides, we observed that even a small model such as Llama3- 8B in its instruct fine-tuned version is mostly able to extract the demanded information from the bunch of unstructured text. We show an excerpt of the dataset in 7. 3.2. Transfer learning 3.2.1 Model architecture As we already outlined in Sec. 2.3, we adopt a sim- ple yet effective strategy for transfer learning from CLIP’s in-distribution to our ILID dataset. Within CLIP’s dual- encoder setup, we must utilize a strategy for the image and text stream. Fig. 5 depicts the used model architectures.
538c8a47-e1a5-470e-b698-5748ee193530
While we estimate that the images we want to learn but also infer from show similar characteristics as CLIP’s in- distribution data compared to other fully out-of-distribution image data as in the case of, e.g., PatchCamelyon [41] (s. Sec. 2.2.2), we employ on the image stream only a sim- ple trainable adapter as proposed by [52]. We tuned the mixing coefficient manually; we observed that a low αcan vastly result in overfitting, while a high value does not necessarily increase the performance significantly during cross-validation. That is why we chose a balanced value of α= 0.5. The adapters reduce the feature by 4as proposed in the original paper [52]. We omitted testing prompt tun- ing on the image stream as introduced by APEX [54] since we estimate a relatively low distribution shift from the CLIP dataset to ILID regarding the images.
1f5f327c-74ec-44ab-95d4-8cfae66b27af
In contrast, prompt engineering is a crucial task for learning, as well as inference with textual, promptable mod- els. In a preliminary study, we have already observed that vanilla CLIP performs differently, given different prompt templates like ”a photo of {}. ”compared to ”a photo of{}, an industrial product. ” The difference from the minor change results from the prompts CLIP was pre-trained with, which follow similar characteristics. Having not to dis- cretely prompt-tune manually motivated us to utilize CoOp [53] as a continuous prompt learning method. Besides, we also evaluate in the experimental study (s. Sec. 4) the per- formance of adding an additional adapter to the text stream.
8b3feab4-6369-4f1c-bc12-6199de352bc7
3.2.2 Training During the pre-training of CLIP, a very large minibatch size of 32,768was used, which took for the largest Vision Transformer (ViT) configuration (428M parameters) a to- tal of 12 days on 256 V100 GPUs [23]. Compared to the pre-training, during transfer learning with CoOp, we have a total of cn×512trainable weights ( cn=number of con- text vectors), which can be managed on a single consumer GPU in a reasonable time. However, the batch size has to be chosen wisely from the memory point of view, as well as by looking at the dataset labels. Given 32k samples per minibatch out of a total of 400M, the chance, utilizing random sampling, that non-contrastive samples are included in one minibatch is negligibly slight.
6176c154-1523-4a2a-b1aa-dede6f3b6c30
In contrast, fine-tuning or transfer learning approaches typ- ically contrast all possible class labels against a set of im- ages [52–54, 56] during the benchmark studies on datasets like ImageNet [57], which is why non-contrasting samples are not possible as long as the classes are conceptually far away from each other. The assembled ILID dataset does not have any class concept, meaning that we, as a priori, do not know how two samples and their labels are semanti- cally close to each other. Contrasting a set of images against all possible labels is infeasible memory-wise; that is why we can not follow this training method and only contrast the images and their labels inside one batch as done dur- ing pre-training. This change led us to employ a different optimizer from the one used in the original CoOp imple- mentation since Stochastic Gradient Descent (SGD) would not converge given the smaller batch size.
501e42cb-a4c1-4765-9bfb-b94773f45408
We changed from vanilla SGD to Adadelta [58], an SGD optimizer that adapts learning rates over time using only first-order information. 4. Experiments In this section, we present a series of studies utilizing ILID, designed to evaluate the effectiveness of the dataset and transfer learning approach for different tasks. We begin with the dataset properties (s. Sec 4), describe the exper- imental setup (s. Sec. 4.2), and present quantitative results on cross-validation (s. Sec. 4.3) as well as training and in- ference on a different label type (s. Sec. 4.4). Further, we present the results of a downstream task on segmentation (s. Sec. 4.5). 6
337bb4be-b7a5-40b3-ac5e-3ed9183af4c9
Adapter + 1 - α α Adapter Image Encoder Text Encoder Adapter Prediction „…hinge…“ „…handle…“ „… rod end…“ [P 1 ] [P 2 ] [ P x ] [..] „hinge“ Image Encoder Text Encoder Prediction Learnable Frozen (a) CLIP Adapter (b) CoOp Figure 5. The architectures used in this work: (a) CLIPAdapter [52] and (b) CoOp [53]. 4.1. Dataset For the presented ILID, as of now, we crawled five different online shops, resulting in 12,537valid samples, including a diverse range of products ranging from standard elements small in size like hinges, linear motion elements, bearings, or clamps to larger ones, like scissor lifts, pallet trucks, etc. (an excerpt is depicted in 7).
0549c4e5-cd6e-49f3-b339-ab6d3971b3cc
steel stainlessclampplungerclampingleverball indexingadjustablewith aluminumhandleknob connectorspringhingelatchprofileplastictogglelinearhandswivelgrip bearinggear assemblyvalve handlesstarfeet handwheellevelingscrewlockplaterollersetjoint aluminium103 Figure 6. Top- 40word occurrences in label label short . Fig. 6 depicts the top- 40word occurrences in label la- belshort , showing that typical concepts of industrial stan- dard parts like clamp ,lever ,handle ,knob ,hinge , orswivel are pronounced represented but also material types ( steel, aluminum /aluminium ) and properties ( stainless ) as well. Tab. 1 lists the number of unique labels per label cate- gory and hints at the dataset’s diversity. Obviously, with in- creasing words (on average: label short <label long<de- scription ), the number of label-wise unique labels increases.
bb6c01d7-9118-4e87-8855-d7487f3c3ad3
So, nearly every sample has a unique description , but only two labels, on average, share the same label short . Since we do not account for minor preposition words like a/an/the in the labels, the labels are slightly more equal on the se-mantically level. However, we estimate a good diversity in the dataset, and since we do not account for preposition words in the counting, at least three to four samples are in- cluded per semantical similar class, which should suffice for a tuned CLIP to outperform fully supervised models (s. Sec. 2.2.2). We use the presented version of ILID in the following experiments. Table 1. Number of unique labels per label category. label short label long material material finish description 6785 8476 2899 3375 11452 4.2.
a7b29b64-6277-48a8-917d-eb2ee904c2b1
Setup We build upon the code base of Dassl [59, 60] and trained on a single 4090 GPU. We chose random sampling, an image input size of 224×224, and CLIP’s pre-trained ViT-B/16 as the image encoder instead of the ResNet version, as ViTs have much less image-specific inductive bias than CNNs. We initialized CoOp with a context length of cn= 10 if not otherwise stated. We trained with a batch size of 64while testing with only 32samples. This increases the accuracy during validation/testing compared to training accuracy, but we feel this is more realistic in the case of real-world ap- plications contrasting only 32different conceptual object classes in one application. We applied common data aug- mentation techniques of randomly resizing and cropping as well as flipping edges during training. Besides, we normal- ized ILID’s image data.
8056e75c-3cb2-47c3-89f5-e0214ca27a68
We use Adadelta [58] with a learn- ing rate of 0.15and a cosine learning rate scheduler with a weight decay of 1e-3. Besides, we used 3warm-up epochs with a constant learning rate of 1e-2to prevent rapid param- 7
40c40951-06d9-45f5-9af1-8afcaa0da539
eter changes in the initial training stages, which can lead to early overfitting. 4.3. Quantitative results Since we do not have a different real-world language- image dataset at hand, we used 6-fold cross-validation during the evaluation of the different model architectures. Fig. 7 depicts the validation results of training on the label short and label long with CoOp, CoOp + image adapter (CoOpIA) with αi= 0.5, CoOp + image αi= 0.5 and text adapter αt= 0.2(CoOpIATA), image and text adapter only (CLIPAdapter) (same αs), and the zero-shot CLIP performance. All accuracies are derived from the top-1 predictions. Additionally, we listed the top-3 ac- curacies for zero-shot CLIP and CoOpIATA (dashed lines). A first observation ( ▷Obs.
23165372-5e95-4fa7-84d6-b55040da54f6
1) is that all transfer learning approaches effectively outperform CLIP’s zero-shot capa- bilities, even the top-3 accuracies after training for ≈20 epochs, highlighting that the ILID is out-of-distribution. Even training on the less information-rich label short out- performs CLIP’s zero-shot capabilities. CLIP highly depends on the chosen prompt template (▷Obs. 2). If we look at training on the label long, the zero-shot (prompt template ”a photo of an industrial prod- uct{label long}”) accuracy is lower than all other trained methods initialized with random weights in the adapters and CoOp with ”X X X X a photo of an industrial prod- uct{label long}”, which tokens will be optimized except for the label. CLIP performs poorly if given a prompt that deviates much from the ones during pre-training. That es- pecially accounts for combining the prompt template with multi-word product descriptions.
409cd147-2f6e-4950-ae00-1b4e15ce643a
As expected, the more trainable weights we add, the better the model adapts to the data, while the overall do- main generalization to the in-distribution data achieves in the case of label short andlabel long an accuracy of maxi- mum 79.93% and84.31%, respectively, an image adapter is crucial to effective transfer learning in this case ( ▷Obs. 3). Moreover, the top-performing model also depends on prompt learning, but adapter-styled tuning performs better than only prompt learning on the text stream ( ▷Obs. 4). However, adapting to images will reduce the model’s per- formance on slight to out-of-distribution data. That means inference on images that vastly differ from catalog-style ones will definitely have lower performance than the in- distribution images. However, the trained model will still outperform zero-shot CLIP, as we will see in the following sections.
93a6d58b-205f-4186-a25d-6a872dfa52be
To gain an understanding of how transfer learning af- fects the embeddings further, we derived the image and text embeddings after training on the full ILID given the label label short for100 epochs. Fig. 8 visualizes thehigh-dimensional embeddings of the same 100 samples. With each transfer learning method, adding more trainable weights, the text and image embeddings more jointly share the same embedding space. Further, multiple text embed- dings get so close that they are hard to distinguish in the t-SNE diagram at all, which we estimate follows that the transfer learning approaches learn to group semantically close concepts, while in the zero-shot case, these are still more widely clustered. Moreover, image and text embed- dings are much more pronounced after transfer learning than in the zero-shot case. 4.4.
5b9e24eb-f3a0-4c94-a429-f92a8be0d001
Prompting for materials Besides training and testing on the label short and la- bellong, we additionally trained CoOpIATA for 100epochs on the material label with the initial prompt ”X X X X a photo of an industrial product with material {}”. We then evaluated the zero-shot and CoOpIATA performance on the images depicted in Fig. 9 while choosing for the zero-shot test a prompt template similar to for training CoOpIATA. The results are listed in Tab. 2. Surprisingly, CLIP’s zero-shot performance shows 2out of5true positives, while the transfer learning result in 5out of5. Further, looking at the scores, we see that our proposed transfer learning method produces much higher confidence in every case, which follows that the different concepts of materials are not in-distribution in the zero-shot case.
de75c89f-f3a9-454a-8a17-9d66c4efac40
In- terestingly, a prompt including {aluminum }results in lower scores than using the word {aluminium }, which points out that the subtleties or discrepancies of the language used in an industrial context are not mapped after the transfer learn- ing nor in the zero-shot case. That is why we added both words in the prompts. Further, after transfer learning, judg- ing based on the scores, there is still slight confusion be- tween the material concepts of aluminum and polyamide as well as polyamide and brass. We estimate that the trans- fer learning introduced a specific object-material-awareness but is still heavily influenced by other image characteristics, like, in our case, the yellow taint. The given task might not serve a real-world industrial vi- sion use case at this stage, but it shows how ILID can serve different tasks at hand by combining images with different (broad) language information during training.
6febc2bb-52b7-4d86-a084-b70a1aea7cdd
These results again underline a natural language supervised VFM’s rich multimodal capabilities. 4.5. Language-guided segmentation A typical downstream task is a language-guided segmen- tation utilizing the Segment Anything Model (SAM) [61]. SAM is a class-agnostic point promptable image segmen- tation model that outputs hierarchical masks and predicted Intersection over Unions (IoU). Without the need for man- ual intervention, an automatic mask generation pipeline can 8
80760d96-3b98-4193-8dd1-95de3b02da0b
10 20 30 40 50 60 70 80 90 100 Epoch2030405060708090x-val/acc (%)(a) Label Short CoOp (top-1) CoOpIA (top-1) CoOpIATA (top-1/3) CLIPAdapter (top-1) Zero-shot CLIP (top-1/3) 10 20 30 40 50 60 70 80 90 100 Epoch2030405060708090x-val/acc (%)(b) Label LongFigure 7. Results of 6-fold cross-validation during transfer learning using different approaches on the ILID. Zero-shot CLIP Image embeddings T ext embeddings CoOpIA CoOpIATA Figure 8. t-SNE diagrams from the same randomly selected 100samples (CoOpIA and CoOpIATA were trained for 100epochs on the full ILID given the label label short ).
65434ad0-b437-4657-a987-cacb6c2822f1
sample a point grid and subsequently use Non-Maximum Suppression (NMS) to diminish through merging a large set of masks to form more precise proposals. In the sim- plest form, language-guided image segmentation based on SAM and CLIP can be employed by applying CLIP onto all generated masks, which we cut out with a particular delin- eation factor. CLIP’s softmaxed logits can then be thresh- olded to get the final per-mask class-wise predictions. We only contrasted the object to prompt against an empty class label. Contrasting only two prompts is challenging since the model’s overconfidence in one of them is the most pro- nounced. We chose to do so to avoid any bias by introducing hard negative prompts. For the language-guided segmentation, we used a CoOp- IATA model trained on the complete ILID dataset given the label long for40epochs.
e61413d8-e564-4d8c-8155-334851f1a586
For completeness, it should be mentioned that we did not compare it against the other ap- proaches, e.g., CLIPAdapter. Fig. 10 depicts the segmentation results in a challenging scene composed of multiple collets stacked on a trolley. The zero-shot results do have many true positives, but overall, we are not able to observe any further prediction patterns.In contrast, the transfer learning approach can effectively distinguish between a mask containing a collet and a mask that does not. With only 17word occurrences of ”collet” in ILID’s label long labels, the resulting model’s confidence compared to zero-shot CLIP effectively demonstrates the proposed method. Additionally, the images relating to the labels do not contain collets of the same shapes and sizes, which emphasizes CLIP’s learned rich representations. We discuss two further examples in 8. 5.
232e949e-3543-421e-8ba4-a5194f23808a
Conclusion and Outlook Using VFMs as a building block in an industrial vision ap- plication is a promising and transforming technique, im- proving systems’ accuracy, speed, and reliability, e.g., in- volved in inspection, robotic control, parts identification, and process control, leading to enhanced operational effi- ciencies and product quality. As we outlined in Sec. 2.1, up to this date, literature only has a limited number of use case ideas regarding using VFMs in industrial applications, which we want to motivate further. This work strived to make a step towards enabling em- ploying VFM in industrial machine vision applications by 9
502cc711-e1ad-46d0-90c9-68e606cfdb1e
(a) (b) (c) (d) (e) Figure 9. Five different real-world images used for prompting material properties. Input Zero-shot CLIP Ours Figure 10. Language-guided segmentation results given prompt ”collet” compared to zero-shot CLIP under the same settings (segmentation properties and thresholds). Table 2. Scores on predicting the object’s material properties in the images from Fig. 9 (bold indicates the highest scores; underlined values correspond to the ground truth). (a) (b) (c) (d) (e) Zero-shot CLIP ”steel” 0.024 0.113 0.330 0.168 0.059 ”polyamide” 0.149 0.196 0.062 0.107 0.208 ”thermoplastic” 0.245 0.141 0.050 0.034 0.
b2896e49-1897-42de-bfa3-3687981694b2
097 ”aluminum or aluminium” 0.043 0.143 0.166 0.238 0.094 ”anodized aluminum or 0.030 0.143 0.070 0.064 0.023 aluminium” ”plastic” 0.352 0.244 0.099 0.107 0.280 ”brass” 0.156 0.020 0.223 0.282 0.240 CoOpIATA trained on the material label ”steel” 0.007 0.033 0.950 0.829 0.137 ”polyamide” 0.135 0.368 0.004 0.008 0.361 ”thermoplastic” 0.010 0.004 0.002 0.001 0.160 ”aluminum or aluminium” 0.009 0.085 0.
39ce2dc5-052b-4398-b094-3eeeb721dacb
020 0.011 0.001 ”anodized aluminum or 0.007 0.374 0.003 0.007 0.001 aluminium” ”plastic” 0.694 0.135 0.008 0.041 0.077 ”brass” 0.139 0.000 0.012 0.104 0.264 introducing the Industrial Language-Image Dataset (ILID) to bring industrial context into CLIP and evaluating ef- fective self-supervised transfer learning from the dataset. We demonstrated this by evaluating downstream tasks from prompting for material properties to language-guided seg- mentation. With only a limited dataset size of ≈12ksam-ples, the results show promising opportunities in machine vision applications when increasing the dataset size or fur- ther restricting it to more specific domains.
618b1138-0a2c-4b55-bb2a-bc72dca91dad
One can argue that the bigger digital giants like OpenAI or Meta can also incorporate industrial data during the train- ing of their models; however, the overall proposed method from dataset curation to fine-tuning CLIP also suits, e.g., companies with intellectual property constraints or limita- tions in available computing resources in employing VFMs. Nevertheless, fine-tuning expert models for specific tasks is a common step in creating an AI application, which we, e.g., showcased, given the transfer learning from material properties. Future work must also elaborate on training with ILID’s other labels, like description , to further discuss op- portunities for other applications. The current limitations we observed on the text stream are especially the limited learned language subtleties and discrepancies as they occur in industrial contexts.
bcb5b328-fb24-4104-97d8-fe4dcf0e0875
The con- fusion between the same concept but differently termed in American (aluminium) and British (aluminum) English shows that there is a need for pre-training of the text encoder with broader natural language, e.g., even with extended con- text, which would enable not only training on shorter image labels. Further, on the image stream, we observed that the model generalizes well to a variety of an object’s different views but does less perform well when contrasting between finer-grained different object types. Here, a custom expert model is probably more suited than transfer learning from a dataset that includes many different object concepts. The 10
3dba94ab-c179-40e0-bf74-4b3bfc2ab21c
most limiting characteristic is including or inferencing with dimensional quantities, which can hardly be solved when training on images captured with different cameras and their individual intrinsics. With this work, we hope to encourage the industrial com- munity to employ and work on using VFM in the industrial domain more and more. Therefore, we publicly provide ILID and the code used during training. In the future, we plan to continue increasing the dataset size by incorporat- ing more web catalogs. Acknowledgments This work is part of the research project Intelligent Digital Cabin Twin (InDiCaT) under the grant number 20D1902C, supported by the Federal Ministry for Economic Affairs and Climate Action (BMWK) as part of the Federal Aeronautical Research Programme LuFo VI-1 . We thank M ¨ADLER GmbH for granting us the rights to use some of their product images (included in Fig.
6d36576a-041a-4ef7-9743-ac41309a839f
2, 3, 5, 11, and 12) in this publication. CRediT author statement K. Moenck: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Resources, Data Curation, Writing – original draft, Writing - review & editing, Visualization, Supervision, Project administration; D.T. Thieu: Conceptualization, Methodology, Software, Formal analysis, Investigation, Data Curation, Writing – original draft; J. Koch: Writing - review & editing; T. Sch¨uppstuhl: Supervision, Project administration, Funding acquisition, Writing - review & editing. References [1] D. Schoepflin, D. Holst, M. Gomse, T. Sch ¨uppstuhl, Syn- thetic training data generation for visual object identifica- tion on load carriers, Procedia CIRP 104 (2021) 1257–1262.
b76eaa09-c104-4621-bd39-252dffdf7d7f
doi:10.1016/j.procir.2021.11.211 . 1 [2] D. Schoepflin, K. Iyer, M. Gomse, T. Sch ¨uppstuhl, Towards synthetic ai training data for image classification in intralo- gistic settings, in: Sch ¨uppstuhl (Ed.) 2022 – Annals of Sci- entific Society, Springer Cham, 2022, pp. 325–336. doi: 10.1007/978-3-030-74032-0_27 . [3] D. Holst, D. Schoepflin, T. Sch ¨uppstuhl, Generation of syn- thetic ai training data for robotic grasp-candidate identifica- tion and evaluation in intralogistics bin-picking scenarios, in: K.-Y . Kim (Ed.
e4e5881a-3b57-4a76-910b-df1c5db494b4
), Flexible Automation and Intelligent Man- ufacturing, Lecture Notes in Mechanical Engineering Ser, Springer International Publishing AG, Cham, 2022, pp. 284– 292.doi:10.1007/978-3-031-18326-3_28 . [4] O. Schmedemann, M. Baaß, D. Schoepflin, T. Sch ¨uppstuhl, Procedural synthetic training data generation for ai-based de- fect detection in industrial surface inspection, Procedia CIRP 107 (2022) 1101–1106. doi:10.1016/j.procir. 2022.05.115 .
836bee01-b230-41d7-816a-f29142e42f08
1[5] B. Drost, M. Ulrich, P. Bergmann, P. Hartinger, C. Steger, Introducing mvtec itodd — a dataset for 3d object recog- nition in industry, in: 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), IEEE, 2017, pp. 2200–2208. doi:10.1109/ICCVW.2017.257 . 1 [6] P. Bergmann, M. Fauser, D. Sattlegger, C. Steger, Mvtec ad — a comprehensive real-world dataset for unsupervised anomaly detection, in: 2019 IEEECVF Conference on Com- puter Vision and Pattern Recognition, IEEE, Piscataway, NJ, 2019, pp. 9584–9592. doi:10.1109/CVPR.2019. 00982 .
f33a3f83-6de4-44b1-a116-538b91b6a8b3
[7] P. Bergmann, K. Batzner, M. Fauser, D. Sattlegger, C. Ste- ger, The mvtec anomaly detection dataset: A comprehensive real-world dataset for unsupervised anomaly detection, In- ternational Journal of Computer Vision 129 (4) (2021) 1038– 1059. doi:10.1007/s11263-020-01400-4 . [8] P. Bergmann, K. Batzner, M. Fauser, D. Sattlegger, C. Ste- ger, Beyond dents and scratches: Logical constraints in un- supervised anomaly detection and localization, International Journal of Computer Vision 130 (4) (2022) 947–969. doi: 10.1007/s11263-022-01578-9 .
a57d219e-3d0f-44f0-857a-0058d21d3fa2
1 [9] H. Bai, S. Mou, T. Likhomanenko, R. G. Cinbis, O. Tuzel, P. Huang, J. Shan, J. Shi, M. Cao, Vision datasets: A bench- mark for vision-based industrial inspection (2023). doi: 10.48550/arXiv.2306.07890 . 1 [10] L. B ¨usch, J. Koch, D. Schoepflin, M. Schulze, T. Sch ¨uppstuhl, Towards recognition of human actions in collaborative tasks with robots: Extending action recognition with tool recognition methods, Sensors (Basel, Switzerland) 23 (12) (2023). doi:10.3390/s23125718 .
7c93d51f-228c-4b22-af93-35a69e079355
1 [11] J. Zhang, J. Huang, S. Jin, S. Lu, Vision-language mod- els for vision tasks: A survey, IEEE transactions on pat- tern analysis and machine intelligence PP (2024). doi: 10.1109/TPAMI.2024.3369699 . 1 [12] R. Bommasani, D. A. Hudson, E. Adeli, R. Altman, S. Arora, S. von Arx, M. S. Bernstein, J. Bohg, A. Bosselut, E. Brun- skill, et al., On the opportunities and risks of foundation models (2021). doi:10.48550/arXiv.2108.07258 . 1 [13] J. Devlin, M.-w. Chang, K. Lee, K. Toutanova, Bert: Pre- training of deep bidirectional transformers for language un- derstanding (2018).
9034e9b1-8472-4d09-bf93-953286f44b32
doi:10.48550/arXiv.1810. 04805 . 1 [14] P. Budzianowski, I. Vuli ´c, Hello, it’s gpt-2 – how can i help you? towards the use of pretrained language models for task-oriented dialogue systems (2019). doi:10.48550/ arXiv.1907.05774 . 1 [15] T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, et al., Language models are few-shot learners (2020). doi: 10.48550/arXiv.2005.14165 .
a5b7dc71-e655-4074-9a48-47548e2d008a
[16] J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anad- kat, et al., Gpt-4 technical report (2023). doi:10.48550/ arXiv.2303.08774 . 1 [17] H. Touvron, T. Lavril, G. Izacard, X. Martinet, M.-A. Lachaux, T. Lacroix, B. Rozi `ere, N. Goyal, E. Hambro, 11
770a1100-949f-4c87-8925-0d1fcc643f47
F. Azhar, et al., Llama: Open and efficient foundation lan- guage models (2023). doi:10.48550/arXiv.2302. 13971 . 1 [18] H. Touvron, L. Martin, K. Stone, P. Albert, A. Almahairi, Y . Babaei, N. Bashlykov, S. Batra, P. Bhargava, S. Bhosale, et al., Llama 2: Open foundation and fine-tuned chat models (2023). doi:10.48550/arXiv.2307.09288 . 14 [19] M. AI, Introducing meta llama 3: The most capable openly available llm to date (26.05.2024).
763e8fcd-cc58-4f20-ae7e-a894f0e41a38
URL https://ai.meta.com/blog/meta-llama- 3/1, 5 [20] M. Awais, M. Naseer, S. Khan, R. M. Anwer, H. Cholakkal, M. Shah, M.-H. Yang, F. S. Khan, Foundational models defining a new era in vision: A survey and outlook (2023). doi:10.48550/arXiv.2307.13721 . 1 [21] C. Schuhmann, R. Beaumont, R. Vencu, C. Gordon, R. Wightman, M. Cherti, T. Coombes, A. Katta, C. Mullis, M. Wortsman, et al., Laion-5b: An open large-scale dataset for training next generation image-text models (2022). doi: 10.48550/arXiv.2210.08402 .
47d6ea10-0f26-4c5b-a192-03741ae7f5d1
1, 2 [22] S. Changpinyo, P. Sharma, N. Ding, R. Soricut, Conceptual 12m: Pushing web-scale image-text pre-training to recog- nize long-tail visual concepts (2021). doi:10.48550/ arXiv.2102.08981 . 1, 2, 5 [23] A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al., Learning transferable visual models from natural language supervision (2021). doi:10.48550/arXiv.2103. 00020 . 1, 4, 6 [24] C. Jia, Y . Yang, Y . Xia, Y .-T.
b3390305-d11b-46b3-aee8-852376d0a028
Chen, Z. Parekh, H. Pham, Q. Le V , Y . Sung, Z. Li, T. Duerig, Scaling up visual and vision-language representation learning with noisy text su- pervision, International Conference on Machine Learning (2021). doi:10.48550/arXiv.2102.05918 . 1 [25] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition (2015). doi:10.48550/arXiv. 1512.03385 . 2 [26] A. Hornberg, Handbook of machine and computer vision: The guide for developers and users, second, revised and up- dated edition Edition, Wiley-VCH, Weinheim, 2017. doi: 10.1002/9783527413409 .
d35886ed-fe60-44a0-87f0-6e15cd69da2d
2 [27] A. Naumann, F. Hertlein, L. D ¨orr, S. Thoma, K. Furmans, Literature review: Computer vision applications in trans- portation logistics and warehousing (2023). doi:10. 48550/arXiv.2304.06009 . 2 [28] K. Moenck, A. Wendt, P. Pr ¨unte, J. Koch, A. Sahrhage, J. Gierecker, O. Schmedemann, F. K ¨ahler, D. Holst, M. Gomse, et al., Industrial segment anything – a case study in aircraft manufacturing, intralogistics, maintenance, repair, and overhaul (2023). doi:10.48550/arXiv.2307. 12674 . 2, 3 [29] J. Wang, Y . Tian, Y .
459277bc-9089-4d4c-abd3-b89a57ebc086
Wang, J. Yang, X. Wang, S. Wang, O. Kwan, A framework and operational procedures for metaverses-based industrial foundation models, IEEE Trans- actions on Systems, Man, and Cybernetics: Systems 53 (4) (2023) 2037–2046. doi:10.1109/TSMC.2022. 3226755 . 3[30] H. Zhang, S. S. Dereck, Z. Wang, X. Lv, K. Xu, L. Wu, Y . Jia, J. Wu, Z. Long, W. Liang, et al., Large scale foun- dation models for intelligent manufacturing applications: A survey (2023). doi:10.48550/arXiv.2312.06718 .
44643e78-7005-4ac9-9550-3a06d7664f8c
3 [31] L. Makatura, M. Foshey, B. Wang, F. H ¨ahnLein, P. Ma, B. Deng, M. Tjandrasuwita, A. Spielberg, C. E. Owens, P. Y . Chen, et al., How can large language models help humans in design and manufacturing? (2023). doi:10.48550/ arXiv.2307.14377 . 3 [32] C. Picard, K. M. Edwards, A. C. Doris, B. Man, G. Gian- none, M. F. Alam, F. Ahmed, From concept to manufactur- ing: Evaluating vision-language models for engineering de- sign (2023). doi:10.48550/arXiv.2311.12668 . 3 [33] Y .
48913af9-3005-4e19-bb09-3a6c6c7fa73c
Mori, H. Takahashi, R. Oka, Image-to-word transforma- tion based on dividing and vector quantizing images with words, in: First international workshop on multimedia in- telligent storage and retrieval management, V ol. 2, 1999. 3 [34] A. Quattoni, M. Collins, T. Darrell, Learning visual rep- resentations using images with captions, in: IEEE Confer- ence on Computer Vision and Pattern Recognition, 2007, IEEE Computer Society, Los Alamitos, Calif., 2007, pp. 1–8. doi:10.1109/CVPR.2007.383173 . 3 [35] K. Desai, J. Johnson, Virtex: Learning visual representa- tions from textual annotations (2020). doi:10.48550/ arXiv.2006.06666 .
b7ca61c3-ea44-499f-b1ea-650aa6f92232
3 [36] M. B. Sariyildiz, J. Perez, D. Larlus, Learning visual rep- resentations with caption annotations (2020). doi:10. 48550/arXiv.2008.01392 . 3 [37] Y . Zhang, H. Jiang, Y . Miura, C. D. Manning, C. P. Langlotz, Contrastive learning of medical visual representations from paired images and text (2020). doi:10.48550/arXiv. 2010.00747 . 3 [38] S. Peng, K. Genova, C. Jiang, A. Tagliasacchi, M. Polle- feys, T. Funkhouser, Openscene: 3d scene understanding with open vocabularies (2022). doi:10.48550/arXiv. 2211.15654 .
eaf4e12f-26fb-4404-80f1-ae1309a34590
3 [39] P. Helber, B. Bischke, A. Dengel, D. Borth, Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification (2017). doi:10.48550/arXiv. 1709.00029 . 4 [40] G. Cheng, J. Han, X. Lu, Remote sensing image scene clas- sification: Benchmark and state of the art, Proceedings of the IEEE 105 (10) (2017) 1865–1883. doi:10.1109/ JPROC.2017.2675998 . 4 [41] B. S. Veeling, J. Linmans, J. Winkens, T. Cohen, M. Welling, Rotation equivariant cnns for digital pathology (2018). doi:10.48550/arXiv.1806.03962 .
4afb6109-0b8e-4d2f-a02e-a46ef6aa886c
4, 6 [42] A. Kolesnikov, L. Beyer, X. Zhai, J. Puigcerver, J. Yung, S. Gelly, N. Houlsby, Big transfer (bit): General visual rep- resentation learning (2019). doi:10.48550/arXiv. 1912.11370 . 4 [43] T. Chen, S. Kornblith, K. Swersky, M. Norouzi, G. Hinton, Big self-supervised models are strong semi-supervised learn- ers (2020). doi:10.48550/arXiv.2006.10029 . 4 [44] Y . Li, F. Liang, L. Zhao, Y . Cui, W. Ouyang, J. Shao, F. Yu, J. Yan, Supervision exists everywhere: A data efficient con- 12
e3c3b183-f591-460b-9c5f-612f03b6cab9
trastive language-image pre-training paradigm (2021). doi: 10.48550/arXiv.2110.05208 . 4 [45] S. Goel, H. Bansal, S. Bhatia, R. A. Rossi, V . Vinay, A. Grover, Cyclip: Cyclic contrastive language-image pre- training (2022). doi : 10 . 48550 / arXiv . 2205 . 14459 . [46] X. Hu, K. Zhang, L. Xia, A. Chen, J. Luo, Y . Sun, K. Wang, N. Qiao, X. Zeng, M. Sun, et al., Reclip: Refine contrastive language image pre-training with source free domain adap- tation (2023). doi:10.48550/arXiv.2308.03793 . 4 [47] J. Yu, Z. Wang, V .
b3092cc1-6c8b-472e-b10a-3474f253675e
Vasudevan, L. Yeung, M. Seyedhosseini, Y . Wu, Coca: Contrastive captioners are image-text foun- dation models (2022). doi:10.48550/arXiv.2205. 01917 . 4 [48] Y . Rao, W. Zhao, G. Chen, Y . Tang, Z. Zhu, G. Huang, J. Zhou, J. Lu, Denseclip: Language-guided dense prediction with context-aware prompting (2021). doi:10.48550/ arXiv.2112.01518 . [49] L. Yao, R. Huang, L. Hou, G. Lu, M. Niu, H. Xu, X. Liang, Z. Li, X. Jiang, C. Xu, Filip: Fine-grained interactive language-image pre-training (2021).
a8240f23-30ad-4c44-ba62-e08f33a153a6
doi:10.48550/ arXiv.2111.07783 . 4 [50] N. Mu, A. Kirillov, D. Wagner, S. Xie, Slip: Self-supervision meets language-image pre-training (2021). doi:10. 48550/arXiv.2112.12750 . 4 [51] Q. Sun, Y . Fang, L. Wu, X. Wang, Y . Cao, Eva-clip: Im- proved training techniques for clip at scale (2023). doi: 10.48550/arXiv.2303.15389 . 4 [52] P. Gao, S. Geng, R. Zhang, T. Ma, R. Fang, Y . Zhang, H. Li, Y . Qiao, Clip-adapter: Better vision-language models with feature adapters (2021). doi:10.48550/arXiv.2110.
0937333d-8249-48d3-9045-30e46ac3b004
04544 . 5, 6, 7 [53] K. Zhou, J. Yang, C. C. Loy, Z. Liu, Learning to prompt for vision-language models (2022). doi:10.1007/ s11263-022-01653-1 . 5, 6, 7 [54] Y . Yang, J. Ko, S.-Y . Yun, Improving adaptability and generalizability of efficient transfer learning for vision- language models (2023). doi:10.48550/arXiv. 2311.15569v1 . 5, 6 [55] B. Lester, R. Al-Rfou, N. Constant, The power of scale for parameter-efficient prompt tuning (2021). doi:10. 48550/arXiv.2104.08691 .
f2a35e0f-5f86-427e-b745-4f3ca637cfee
5 [56] K. Zhou, J. Yang, C. C. Loy, Z. Liu, Conditional prompt learning for vision-language models (2022). doi:10. 48550/arXiv.2203.05557 . 6 [57] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, Imagenet: A large-scale hierarchical image database, in: IEEE Conference on Computer Vision and Pattern Recog- nition, 2009, IEEE, Piscataway, NJ, 2009, pp. 248–255. doi:10.1109/CVPR.2009.5206848 . 6 [58] M. D. Zeiler, Adadelta: An adaptive learning rate method (2012). doi:10.48550/arXiv.1212.5701 .
c0d895ae-3d10-450e-81c1-6fa0850650de
6, 7 [59] K. Zhou, Y . Yang, Y . Qiao, T. Xiang, Domain adaptive en- semble learning (2021). doi:10.1109/TIP.2021. 3112012 . 7[60] K. Zhou, Z. Liu, Y . Qiao, T. Xiang, C. C. Loy, Domain gen- eralization: A survey (2023). doi:10.1109/TPAMI. 2022.3195549 . 7 [61] A. Kirillov, E. Mintun, N. Ravi, H. Mao, C. Rolland, L. Gustafson, T. Xiao, S. Whitehead, A. C. Berg, W.-Y . Lo, et al., Segment anything (2023). doi:10.48550/ arXiv.2304.02643 .
ff10f718-4361-40c8-84fe-03f5bfb3ba09
8 [62] H. Wang, P. K. A. Vasu, F. Faghri, R. Vemulapalli, M. Fara- jtabar, S. Mehta, M. Rastegari, O. Tuzel, H. Pouransari, Sam- clip: Merging vision foundation models towards seman- tic and spatial understanding (2023). doi:10.48550/ arXiv.2310.15308 . 14 13
59e422f0-5a1d-4530-8101-a848632edf1f
6. Llama-3 prompt We followed basic prompt assembly as described for Llama-2 [18] because up to the date of this publication, there has still been an in-depth explanation of Llama-3 missing. The Llama-2 chat version was trained with a va- riety of system prompts following patterns like ”You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. ” , we also included a similar one but tried to include the targeted domain. The brackets {{}} point out where we insert the data. Listing 1. System prompt used in the ILID generation pipeline’s text transformation step. You a r e a h e l p f u l a s s i s t a n t f o r a company t h a t s e l l s i n d u s t r i a l p r o d u c t s .
8a35fc0b-1b11-4724-ab0a-56272e090675
\n Do n o t ask f o r f u r t h e r d e t a i l s or s t a t e a d d i t i o n a l q u e s t i o n s . \n Do n o t add a d d i t i o n a l i n f o r m a t i o n or d e t a i l s t h a t a r e n o t g i v e n by t h e u s e r . \n Listing 2. User prompt used in the ILID generation pipeline’s text transformation step.
019286ba-0536-437e-a264-caa82dda1922
Summarize ’ Label : {{}} Text : {{}} ’\n r e t u r n i n g t h e f o l l o w i n g i n f o r m a t i o n : \n ( 1 ) a lo ng l a b e l or name of t h e p r o d u c t w i t h o u t i d s , numbers , codes , or s i z e s ( 2 ) a s h o r t l a b e l or name of t h e p r o d u c t w it h a maximum of 4 words and s h o r t e r t h a n t h e l on g l a b e l ( 3 ) d e s c r i p t i o n of t h e p r o d u c t w ith a maximum of 20 words w i t h o u t i d s , numbers , codes ,
b35c4027-8c31-4c57-97a3-21e8660cdb0e
or s i z e s ( 4 ) m a t e r i a l wi th a maximum of 5 words ( 5 ) m a t e r i a l f i n i s h / c o l o r w ith a maximum of 5 words 7. Excerpt from the dataset Fig. 11 and Fig. 12 depict each two samples from the ILID given the keywords ”hinge” and”locking assembly” . Based on the language label, we can observe that the LLM performs differently in extracting the relevant information. As an example, material andmaterial finish confusion oc- curs when the product page states more than one exact prod- uct configuration. 8. Additional language-guided segmentation results Fig. 13 and 14 show supplementary results on language- guided image segmentation. In Fig.
5c75097a-0975-4785-be35-fb32e69136f3
13, we prompted for”socket” , whereas zero-shot CLIP does not predict any mask as positive, while our approach segments all sockets. In Fig. 14, the results of our most challenging scene are depicted, in which we prompt for ”bracket for construction profile” .
22cea8fb-58d1-4306-8a2e-537883a30c34
The brackets are imaged far differently than the ones from catalog images, and sometimes they are barely { "id": "...", "image": "...", "label_short": "clevis mounting hinge", "label_long": "bracket hinge for clevis mounting", "description": "Rigid hinge for clevis mounting applications", "material": "rigid metal", "material_finish": "black oxide finish" } { "id": "...", "image": "...", "label_short": "adjustable hinge mechanism", "label_long": "hinge with adjustable friction mechanism", "description": "A hinge with adjustable friction for smooth motion and secure locking", "material": "zinc die casting", "material_finish": "silver plastic coating" }Figure 11. Two samples from the ILID given the keyword ”hinge” .
98de050c-0103-429f-a73a-4306f28510cd
{ "id": "...", "image": "...", "label_short": "locking qpq assembly", "label_long": "locking assembly with qpq coating", "description": "High corrosion resistance and improved fatigue strength for food safe applications", "material": "steel", "material_finish": "qpq-coated" } { "id": "...", "image": "...", "label_short": "stainless steel locking bar", "label_long": "locking assembly stainless steel bar", "description": "A locking assembly designed for industrial use made from high-quality stainless steel for durability and resistance", "material": "stainless steel aisi", "material_finish": "stainless steel finish" } Figure 12. Two samples from the ILID given the keyword ”locking assembly” . visible.
16b66c62-6e31-4d8c-8910-cb58586b595b
At first sight, the results do not show good perfor- mance, especially since we have a few non-detected brack- ets and a few false positive predictions. We explain the false positive on the top with the cropping strategy, while we have no explanation for the false predictions on the lower right. The false positives can result from the axis-preserving cropping strategy of the used method, in which a cropped segment includes parts of the surroundings. A lot of the false positive segments contain parts of brackets. Employ- ing more sophisticated language-image segmentation meth- ods, like [62], based on CLIP and SAM that do not rely on such a straightforward cropping strategy could prevent such wrongful predictions. In contrast, we observed less perfor- mance during the segment classification with CLIP when the background was not included in the segments. 14
0aa5e298-dcff-4b44-80a3-1cb0265652f2
Input Zero-shot CLIP OursFigure 13. Language-guided segmentation results given the prompt ”socket” compared to zero-shot CLIP under the same set- tings. Input Zero-shot CLIP OursFigure 14. Language-guided segmentation results given the prompt ”bracket for construction profile” compared to zero-shot CLIP under the same settings. 15

This dataset was created using Corpus Creator. This dataset was created by paring a corpus of texts into chunks of sentences using Llama Index.

Downloads last month
3
Edit dataset card