The search functionality is under construction.
The search functionality is under construction.

Open Access
Enhanced Data Transfer Cooperating with Artificial Triplets for Scene Graph Generation

KuanChao CHU, Satoshi YAMAZAKI, Hideki NAKAYAMA

  • Full Text Views

    110

  • Cite this
  • Free PDF (5.6MB)

Summary :

This work focuses on training dataset enhancement of informative relational triplets for Scene Graph Generation (SGG). Due to the lack of effective supervision, the current SGG model predictions perform poorly for informative relational triplets with inadequate training samples. Therefore, we propose two novel training dataset enhancement modules: Feature Space Triplet Augmentation (FSTA) and Soft Transfer. FSTA leverages a feature generator trained to generate representations of an object in relational triplets. The biased prediction based sampling in FSTA efficiently augments artificial triplets focusing on the challenging ones. In addition, we introduce Soft Transfer, which assigns soft predicate labels to general relational triplets to make more supervisions for informative predicate classes effectively. Experimental results show that integrating FSTA and Soft Transfer achieve high levels of both Recall and mean Recall in Visual Genome dataset. The mean of Recall and mean Recall is the highest among all the existing model-agnostic methods.

Publication
IEICE TRANSACTIONS on Information Vol.E107-D No.9 pp.1239-1252
Publication Date
2024/09/01
Publicized
2024/04/30
Online ISSN
1745-1361
DOI
10.1587/transinf.2023EDP7228
Type of Manuscript
PAPER
Category
Image Recognition, Computer Vision

1.  Introduction

Scene graphs have emerged as a pivotal representation for detailing semantic information within a visual scene, by specifying relationships between object pairs [1], [2]. This representation enables reasoning about visual content through the encoded spatial and logical details of object instances and their relations. In modern applications, scene graphs have become foundational for high-level visual tasks like activity parsing [3], image retrieval [1], visual understanding [4], and image captioning [5]. This paper delves into the scene graph generation (SGG) task, aiming to predict objects and their relations from visual input.

SGG models encounter two primary challenges when trained on common dataset [6]: first, the distinct long-tailed distribution of relations [7], [8], and second, the ambiguity caused by semantically similar relation classes (e.g., on/on back of/mounted on) [9]-[11]. The latter exacerbates the issue, as instances within a category may be annotated under multiple confusing classes. Such complexities often bias relation predictions in general SGG models, leading to low recall rates for rare predicate classes. While some unbiased SGG methods [7], [12]-[14] have addressed this, they often sacrifices performance on frequent classes. Hence, it is essential to consider these trade-offs to ensure the model performance on the majority of data is not compromised.

Recently, the training data modification approach has shown promising results for training an unbiased SGG model [9], [10]. Two major concepts for the modification are the addition of new predicate labels and reassignment of existing ones, which can efficiently improve rare class performance. We revisit these concepts through IETrans [10], a baseline data modification method. IETrans encompasses two modules: external transfer for label addition and internal transfer for label reassignment. Notably, the external transfer, while leveraging background triplets for augmentation, doesn’t fully exploit the available data. Given the compositional nature of relational triplets, inter-triplet augmentation appears worthwhile. Additionally, predicate reassignments in the internal transfer are not uniformly reliable. A human evaluation study [10] reveals that only 76% of transferred triplets are deemed reliable. The inconsistency in the degree of semantic confusion, even among identical predicates, suggests that an “entire” transfer strategy might not be optimal. Guided by these findings, our system seeks to address these shortcomings by extending the modification concepts in two key ways: improving upon the data addition process and enhancing the reassignment efficiency.

Our method introduces two novel modules: Feature Space Triplet Augmentation (FSTA) and Soft Transfer. FSTA dynamically creates artificial triplets during training. We can construct new data by enumerating triplet combinations subject-predicate-object' and subject'-predicate-object from a sampled mini-batch. Here, \(x'\) denotes data not from the original triplets. These artificial triplets serve to regularize the relation classification module in the SGG model. We undersample the frequent classes in artificial triplets to shape their predicate distribution. Further, a biased prediction-based sampler selects the class label for \(x'\). This design aims to often sample combinations that are hard to be predicted correctly for a biased model. A pre-trained generator synthesizes the corresponding features based on class labels. Besides, Soft Transfer refines label reassignment by implementing an instance-wise ranking and mapping mechanism. We first compute a reliability score for each reassigned sample from biased model predictions, then select low-scoring triplets for Soft Transfer. Subsequently, a non-binary predicate label is calculated by mapping the reliability score, allowing for finer control over semantic confusion by using this label probability instead of an entire reassignment.

FSTA notably boosts performance on rare classes with increased sample quantity and diversity. Conversely, Soft Transfer alleviates performance loss in frequent classes, a typical compromise when elevating rare class performances. In essence, while FSTA contributes to mean recall (mR) gain, Soft Transfer leads to the recall (R) gain. Collectively, these modules bring reduced performance trade-off that shown in the improved overall metrics, F1@K and Avg@K. Our model-agnostic method was evaluated on the VisualGenome dataset [6], using two types of general SGG models: MOTIF [15] and RelDN [16] with IETrans. In the predcls task, our system outperforms the baseline IETrans by a 3.1% and 7.0% relative gain on the F1@100 metric for MOTIF and RelDN, respectively. Figure 1 illustrates the balanced performance of our method.

Fig. 1  Accuracy comparison between FSTA, Soft Transfer, Full, and the baseline IETrans on Motif (\(1^{st}\) row) and RelDN (\(2^{nd}\) row). In the scatter plots (left), a larger dot size and a darker color represent higher F1@100 and AVG@100 scores, respectively. As shown in the bar plots (right), increased scores in the overall metrics (F1@100 and AVG@100) indicate the alleviated performance trade-off in our full method, consisting of two complementary modules.

To sum up, we make the following contributions:

  1. We propose a novel, model-agnostic method for training a R/mR balanced SGG model. It integrates two complementary modules: FSTA and Soft Transfer, which enhance the baseline IETrans.

  2. We conduct extensive experiments and discussions on VisualGenome and demonstrate the effectiveness of our system.

2.  Related Work

2.1  Biased and Unbiased Scene Graph Generation

Scene Graph Generation (SGG) is first proposed as visual relation detection (VRD) [17], where each relation is independently detected, ignoring the rich contextual information. Later studies in SGG utilizes advanced techniques, e.g., message passing [18], recurrent sequential architectures [15] or contrastive learning [16]. However the accuracy of relationship detection is far from satisfaction due to the heavily biased data. Some authors [19], [20] point out that the predictions of current SGG models often collapse to several general and trivial predicate classes. Instead of only focusing on recall metric, hence they propose a new metric named mean recall, which is the average recall of all predicate classes, as the unbiased metric. Efforts towards developing unbiased SGG models have been noted. BGNN and DT2-ACBS [7], [21] proposes sophisticated re-sampling strategy. Some debiasing solutions [10], [12], [22] are categorized as biased-model based strategies that utilize predictions from biased SGG model. Especially, IETrans [10] adopts triplet-level data transfers over the less precise predicate-level manipulation. Our proposed method is inspired from IETrans, and focuses on data augmentation for inadequate training samples.

2.2  Compositional Learning

Recognition-By-Components theory [23] which illustrates that human representations of concepts are decomposable is especially influential in object recognition. Based on the theory, novel concepts from a few samples can be potentially learnable by composing known primitives. Some authors apply the compositional deep representation into few-shot learning for object recognition [24] and Human-Object Interaction (HOI) detection [25]-[27]. Visual compositional learning frameworks [26], [27] proposed for HOI detection compose HOI training samples from image-pairs and fake object representations to solve the open long-tail issue in HOI detection. Our proposed data augmentation method adapts the compositional learning to SGG task. To overcome the biased data issue in SGG, our sampling strategy of composed training samples plays an important role.

3.  Methodology

A scene graph generation (SGG) model predicts a direct graph \(G\) for an input image \(I \in \mathbb{R}^3\). \(G = \lbrace V, E\rbrace\) contains a set of predicted objects \(V = \lbrace(\mathbf{b}_{i}, c_{e_{i}})\}^{N_{V}}_{i=1}\) and a set of predicted relationships \(E = \lbrace(s_{j}, c_{r_{j}}, o_{j})\}^{N_{E}}_{j=1}\). \(\mathbf{b}_{i} \in \mathbb{R}^4\) denotes the position of an object using bounded box coordinates. \(c_{e_{i}} \in \mathcal{C}_{objects}\) and \(c_{r_{j}} \in \mathcal{C}_{relations}\) belong to the known object and relation classes, respectively. \(s_{j} \in V\) and \(o_{j} \in V\) are nodes connected by the relation \(c_{r_{j}}\). Each element in \(E\) can also be depicted as a subject-predicate-object triplet to convey the intrinsic semantic information.

Conceptually, an SGG model can be seen as a sequence of modules that includes an object detection backbone followed by a relation prediction head. The detection backbone first outputs a set of Region of Interest (RoIs) containing the detected object information. These results are then forwarded to the relation prediction head to refine these detections and predict the relation between the RoI pairs. This work mainly focuses on the scenario where a maximum of one predicate, with the highest score, can be predicted between each RoI pair. This principle aligns with the graph constraint mode described in other research.

Figure 2 illustrates the system overview. In our enhanced data modification approach, we further leverage predictions from a pre-trained yet biased SGG model. Section 3.1 gives a brief overview of the baseline modification, IETrans. Section 3.2 details our FSTA module, elucidating a strategy for triplet augmentation in the feature space during the unbiased training phase. Section 3.3 explains Soft Transfer, a method offering precise control over reassigning predicate labels during the pre-processing stage, ensuring better handling of per-sample semantic confusion. Section 3.4 has our implementation details.

Fig. 2  The system overview of our proposed method. The FSTA and Soft Transfer modules are designed to introduce new concepts to enhance the baseline dataset manipulation module, IETrans. Blocks indicated in blue are prepared during the pre-processing stage, whereas the blocks in purple are designated for the unbiased SGG model training stage.

3.1  Preliminary Introduction: IETrans

IETrans constructs a modified training dataset during pre-processing. It has two steps: external transfer and internal transfer. In the external transfer, it acquires new labels from those no-relation object pairs. By ranking these no-relation prediction scores from the biased model, some object pairs are assigned new predicate labels that have the highest probability. This approach, however, may not fully leverage the available data. On the other hand, internal transfer shifts general predicates to informative ones using a ranking and affinity score filtering method through biased prediction results. For example, “man-on-horse” becomes “man-sitting on-horse”. Nevertheless, the level of ambiguity is context-sensitive, and a binary transfer decision might not effectively capture semantic confusion across all samples.

These transfer steps explicitly adjust the balance of the dataset distribution, based on the property that a rare predicate class is often a more informative version of a frequent class. Linking general to informative predicate pairs reduced semantic ambiguity by discovering the confusion in biased model predictions. The raw frequently prior is employed to compensate the largely sacrificed performance on general predicates. We refer readers to the original publication for more details.

3.2  Feature Space Triplet Augmentation

Given the compositional nature of a relation triplet, it’s possible to construct a new sample from multiple existing ones. The interaction between object-predicate representations in the feature space for SGG models is pivotal. Even though they can be combined in various ways―be it addition [16], concatenation [16], or element-wise multiplication [15]―the upstream feature extractor processes the elements in a triplet independently. As such, when the object representation in a triplet is partially changed to form a new semantically reasonable combination, the relation predictor in the relation head should be encouraged to produce similar outputs. This can be represented as:

\[\begin{align} \begin{aligned} & M(F(\mathbf{f}_{s_{i}}, \mathbf{f}_{p_{i}}, \mathbf{f}_{o_{i}}; \theta_{F});\theta_{M}) \\ & \quad \approx M(F(\mathbf{f}_{s_{i}}, \mathbf{f}_{p_{i}}, \mathbf{f}_{o_{j}}; \theta_{F});\theta_{M}) \end{aligned} \tag{1} \end{align}\]

where \((\mathbf{f}_{s_{i}}, \mathbf{f}_{p_{i}}, \mathbf{f}_{o_{i}})\) denotes the subject-relation-object intermediate representations of the \(i^{th}\) sample. Given that \(i\neq j\) and \((c_{s_{i}}, c_{p_{i}}, c_{o_{j}})\) is a semantically reasonable triplet, \(M(\cdot;\theta_{M})\) is the final predicate classification module, and \(F(\cdot;\theta_{F})\) symbolizes the transitional layers in between. In light of this, artificial triplets can serve as augmented data to regularize the relation predictor during training.

We present feature space triplet augmentation (FSTA) via artificial triplets. Compared with generating new image samples, features are more tractable and computationally efficient without using external knowledge [28], [29]. For an input mini-batch of size \(N_{B}\), the detector backbone yields RoIs of varying numbers with their object prediction results. We sample \(N_{t}\) RoI pairs per image from the pool of RoI pairs that have an overlap score exceeding \(s_{iou}\) with any of the ground-truth triplets in the image. Next, we enumerate a set \(\mathcal{T}_{spo'}\) of combinations from these \(N_{B} \times N{t}\) triplets as subject-predicate-object', where object' represents the object features from all other sampled triplets. After eliminating pairs that are absent from the training set label space, the remaining feature combinations―deemed to be reasonable―are forwarded to the same modules in the relation head. The resulting outputs are employed to compute a regularization term \(\alpha\mathcal{L}_{at}\) to foster consistent predicate predictions on artificial triplets. We use the same loss function (i.e., cross entropy) as in the origin relation head for computing \(\mathcal{L}_{at}\). Figure 3 visualizes our approach to building combinations for artificial triplets.

Fig. 3  Building combinations from batch input proposals. Purple box pairs are excluded for low IoU with ground-truth relations and red box pairs are selected as candidates.

Besides generic artificial triplet synthesis, our FSTA module incorporates two novel features: (1) bi-directional resampling, and (2) Model prediction-based class sampler.

The bi-directional resampling further expands the volume of artificial triplets by enumerating subject'-predicate-object into a new set of combinations \(\mathcal{T}_{s'po}\), where sbject' is sourced from other sampled triplets. As \(\mathcal{T}_{s'po} \cap \mathcal{T}_{spo'} = \emptyset\), this enhances the richness and diversity of the artificial triplets. We define an undersampling parameter \(U_{h} \in [0,1]\) to govern the predicate distribution in artificial triplets. For triplets of frequent relations (termed as “head group”), we retain a random \(U_{h}\) fraction of them. This can effectively achieve a distribution shift toward rare relations in artificial triplets. Overall, We combine artificial triplets built from \(\mathcal{T}_{spo'}\) and \(\mathcal{T}_{s'po}\).

Moreover, we propose a new sampler based on biased model predictions (hence, MP-sampler) on training data. It targets to generate suitable object' classes rather than mere swaps. The motivation is straightforward: Combinations that are difficult to be predicted correctly ought to be sampled more frequently. To begin, we enumerate the candidate object' classes as \(\mathcal{O}_{cand}\) from the dataset label space for a given sbject-predicate class label pair, \((c_{s}, c_{p})\). Then, we define a difficulty score function \(d(\cdot)\) which computes the mean score discrepancy between the top-1 prediction and the ground-truth predicate class:

\[\begin{align} \begin{aligned} & d(c_{s}, c_{p}, c_{o_{i}}) = max(l(c_{s}, c_{p}, c_{o_{i}})) \\ & \hphantom{d(c_{s}, c_{p}, c_{o_{i}}) =} - v(l(c_{s}, c_{p}, c_{o_{i}}), c_{o_{i}}) \end{aligned} \tag{2} \end{align}\]

where \(o_{i} \in \mathcal{O}_{cand}\), and \(l(\cdot) \in \mathbb{R}^{\lvert \mathcal{C}_{relations} \rvert}\) returns the average post-softmax predicate prediction vector for the input combination in MP. \(v(l, i)\) obtains the value of \(l\) at element \(i\). If the correct relation for a combination is often mispredicted, the difficulty score is positive; otherwise, it is zero. The MP-sampler then can generate an object' class following the probability:

\[\begin{align} p(c_{o_{i}} | (c_{s}, c_{p})) = \frac{d(c_{s}, c_{p}, c_{o_{i}})}{\sum_{j=1}^{\lvert \mathcal{O}_{cand} \rvert} d(c_{s}, c_{p}, c_{o_{j}})} \tag{3} \end{align}\]

In short, our emphasis primarily rests on those hard-to-predict combinations when building artificial triplets.

With MP-sampler, we use a generator to synthesize features for object', as sampled classes are not assured to align with the classes from the batch’s swapped features. We collect the ground-truth object features to train a conditional-GAN [30]. Following [31], [32], we define its adversarial loss function as:

\[\begin{align} \min_{G}\max_{D} \mathcal{L}_{wgangp} + \beta \mathcal{L}_{cls} + \gamma \mathcal{L}_{recon} \tag{4} \end{align}\]

where \(\mathcal{L}_{cls}\) and \(\mathcal{L}_{recon}\) regularize the generator output via an object classifier and an reconstructor respectively, of both pre-trained on real data. Having the trained generator \(G\), it is capable of yielding synthetic object features with MP-sampler, and we can construct artificial triplets for \(\mathcal{T}_{spo'}\). The GAN model detail can be found in the appendix. Algorithm 1 outlines the complete procedures of FSTA.

3.3  Soft Transfer

The IETrans internal transfer reassigns relation labels from the general (source) ones to the informative (target) ones. However, some transfers are suboptimal: human evaluation deems only 76% of general-informative pairs as “reliable” [10]. While tail performance can benefit from these transfers, the cost of head performance drop is a concern.

A finer control on individual transfers could improve the transfer efficiency and thus alleviate the tail-head performance trade-off. Instead of a complete label transfer from a general (\(p\rightarrow0\)) to an informative predicate (\(p\rightarrow1\)), we propose the Soft Transfer that assigns non-binary probabilities to source and target predicate classes. Soft Transfer consists of two steps. Firstly, we rank all the reassigned pairs using a triplet-wise reliability score, from which pairs are selected for Soft Transfer. Second, a mapping function converts the reliability score into probabilities for source and target labels.

Based on the observation that transfer reliability varies from one combination to another, we define a preliminary function \(r_{int}(\cdot)\) to estimate the degree of reliability. Given an transfer decision list, each list item includes a triplet index \(i\), a source class \(c_{p}^{src,i}\), and a target class \(c_{p}^{tar,i}\). To determine the reliability score, we use the prediction output difference following

\[\begin{align} r_{int}(i) = v(l_{triplet}(i), c_{p}^{tar,i}) -v(l_{triplet}(i), c_{p}^{src,i}) \tag{5} \end{align}\]

where \(l_{triplet}(i) \in \mathbb{R}^{\lvert \mathcal{C}_{relations} \rvert}\) returns the post-softmax model prediction of triplet \(i\), and \(v(l_{triplet}(\cdot), j)\) retrieves the value of \(l_{triplet}(\cdot)\) at class \(j\). We rank the score in ascending order and pick the top \(k_{s}\)% triplets for Soft Transfer while the other remains.

For those triplets with low reliability scores, we consider them as over-transferred. Thus, a positive probability should be assigned to the source class as the ground-truth label instead of zero. While achieving this and ensuring that the sum of the classes for the label is \(1\), we map the reliability scores to values within the range \([0,1]\). Given a mapping function \(Q(\cdot)\), the post-transferred result for triplet \(i\) can be represented as its label probability:

\[\begin{align} label_{i}(c) = \begin{cases} \frac{1}{1+Q(r_{int}(i))},& \text{if } c = c_{p}^{tar,i}\\ \frac{Q(r_{int}(i))}{1+Q(r_{int}(i))},& \text{if } c = c_{p}^{src,i}\\ 0, & \text{otherwise} \end{cases} \tag{6} \end{align}\]

we set \(Q(\cdot)=1-Q'(\cdot)\), where \(Q'(\cdot)\) is a linear min-max scaling for the reliability scores.

Soft Transfer is applied on the original relation loss and needs no changes to it. Table 1 shows an example of post-transferred annotations. In IETrans, predicates of the selected triplets are reassigned to more informative ones (red). Our Soft Transfer evaluates the reliability of these reassigned predicates and converts them to non-binary values (blue).

Table 1  The differences in relation annotation among the raw dataset, the baseline IETrans (excluding External Transfer), and our proposed Soft Transfer for Fig. 5.

3.4  Implementation Details

We build our work upon an open source SGG model implementation1 [34]. We integrate our system into two prevalent SGG model of distinct types: Motif [15] (which employs LSTM) and RelDN [16] (which utilizes CNN, multi-modality fusion, and contrastive losses). These were selected because they represent a variety of design elements commonly found in popular models. We use a ResNet50-FPN [35] Faster-RCNN [36] as the common detector backbone. The detector backbone is pre-trained on VisualGenome [6] and kept frozen. Figure 4 illustrates how to combine our module with these SGG models. We implement IETrans with the default parameters: \(k_{i}=70\) and \(k_{e}=100\). In the FSTA module, \(N_{t}\) is set to 2 for Motif and 5 for RelDN to balance the number of artificial triplets in a mini-batch, considering the smaller batch size for RelDN. We set \(s_{iou}=0.7\)2, \(U_{h}=0.2\) for Motif, and \(s_{iou}=0.5\), \(U_{h}=0.8\) for RelDN. Both models have a loss coefficient of \(\alpha=0.1\). We omit artificial triplets from \(\mathcal{T}_{s'po}\) if their predicates are not in the tail group. For the soft transfer module, we set \(k_{s}\) to 10 for Motif and 70 for RelDN. In the experiments with a “reweighting” setting, only the original loss function is applied with reweighting, while \(\mathcal{L}_{at}\), the loss for FSTA, does not apply as a standalone module. \(k_{s}\) also changes to 30 for Motif and 90 for RelDN.

Fig. 4  The schematic view illustrates the combination of FSTA and SGG models. We visualize only the flow of \(\mathcal{T}_{spo'}\) with red dotted lines for readability. The green dotted line indicates the point at which features are collected in the preparation stage.

Fig. 5  An example training image in VisualGenome.

4.  Experiments

4.1  Dataset and Evaluation Protocol

We evaluated our system on the benchmark VG150 split of the VisualGenome dataset. This dataset consists of 60,784 training images and 26,446 testing images. It contains 150 object classes and 50 relation classes. Following the approach of [7], we sorted the predicates by cardinality, grouping the top 16, middle 17, and bottom 17 into head, body, and tail groups, respectively.

Our analysis focused on standard SGG tasks: predcls, sgcls, and sgdet [15], [16], [19]. These tasks evaluate the model with incrementally higher demands. For instance, “predcls” only assesses the model’s ability in classifying relations using given object locations and categories. In contrast, “sgdet” evaluates both relation classification and object detection simultaneously. Our primary attention was on predcls since our proposed modules target improving predicate classification performance. We used the Recall(R)@K and mean Recall(mR)@K metrics for both full test set and per-class averaged recall evaluations. It is noteworthy that Recall@K is dominated by the performance of the top frequent classes due to the skewed predicate distribution, whereas mean Recall@K treats all classes equally. Given the observed trade-off between Recall@K and mean Recall@K from earlier studies, we also reported the F1@K (their harmonic mean) and Avg(A)@K (their arithmetic mean) [9], [10] as the “overall” metrics in our comprehensive evaluation. All metrics are the higher the better.

4.2  Comparing to Other Methods

We compared our results with IETrans and several other recent model-agnostic SGG methods (first section of Table 2). IETrans serves as the baseline and is currently one of the best model-agnostic methods available.

Table 2  The performance comparison for the predcls task on VG150. Scores for models listed in the first section are cited from their original papers, while models in subsequent sections use our implementation. “Model++X” is shorthand for “Model+IETrans+X”. The best overall scores within each section are highlighted in bold. (Unit: %)

Original baseline and our re-implementation. We compared our method with the reproduced baselines (denoted as 1). For Motif+IETrans in the predcls task, the reproduced version yielded similar scores to those of the original, with a slightly higher R@100 and lower mR@100. These differences may due to some implementation variations in the base SGG model. Therefore, we use the reproduced version as our standard because it maintains identical implementation settings and IETrans transfer lists, consistent with our proposed methods. The original IETrans paper did not present results for RelDN; therefore, we also compared our results with a reproduced version. In summary, all our implementations share the basic settings to ensure a fair comparison.

Improved relation prediction over the baseline. Table 2 summarizes the scores for predcls. The results reveal that our method substantially outperformed the baseline for both the Motif and RelDN models. Specifically, the F1@100 score rose from 41.7 to 43.0 (a 3.1% relative gain) for Motif, and from 35.7 to 38.2 (a 7.0% relative gain) for the RelDN model. The A@100 score also increases. With FSTA, the standout feature was the mR enhancement in tail classes (e.g., from 16.4 to 22.0 for Motif). The artificial triplets generated in FSTA enriched the variation of triplets available for the relation predictor, aiding especially the sparse classes. As for frequent classes, the score decline is minor. On the other hand, Soft Transfer was intended to reduce the degree of label reassignment for less reliable transfers. This led to a score trend the opposite of the original IETrans: while recall scores raised, the tail mean recall scores decreased (e.g., R@100 increases from 57.1 to 60.8 for Motif, and 39.9 to 53.8 for RelDN). In certain cases, Soft Transfer can slightly reduce the F1 score, because the harmonic mean prioritizes enhancements in the smaller one. Nonetheless, the Avg@100 witnessed a notable boost with Soft Transfer. Combining both modules, the full system leveraged their complementary benefits, consistently delivering among the top F1/Avg@100 results for both models, indicating an effective balance of trade-offs.

Compatible with the reweighting setting. We also adhere to the original settings described in the IETrans paper to compare the models when integrated with the “reweighting” technique (+rwt) [10]. Our method proved efficacious even under this setting. Both the FSTA and Soft Transfer modules served their intended purposes, driving improvements across rare and frequent classes alike. The “Motif++Full+rwt” method achieves an increase in F1@100 from 45.1 to 46.1, and in Avg@100 from 46.2 to 47.5, thereby demonstrating the mitigation of the performance trade-off. For RelDN with reweighting, the performance was not as beneficial as for the Motif models. Although the mR@100 for the tail group further increased, the impact on R@K cannot be overlooked, leading to a dip in the overall scores. One possible explanation for this could be the architecture of the RelDN model, which already incorporates a frequency prior branch. Consequently, we did not add the frequency prior values during inference for the RelDN models, while we did so for the Motif models following [10]. This leads to a more serious degradation in head classes, despite having the best performance on tail classes. However, our method still consistently surpassed the baseline in the RelDN +rwt setting.

Similar trends observed for sgcls and sgdet. Table 3 showcases the digested results for sgcls and sgdet. Here, we noticed trends analogous to those in predcls. For RelDN, the full version achieves the best F1@100 and the second-best A@100. For the Motif sgdet, the full version outperforms all the others on both overall metrics (i.e., a 5.7% and 7.4% relative gain for F1@100 and A@100 over the baseline method, respectively). However, the FSTA module yields some unexpected results in the sgcls task. One potential cause is that we applied identical FSTA settings across all tasks. However, sgcls uniquely relies on ground-truth boxes only for input proposals, which is different from the other tasks. This difference might result in distinct regularization effects, as the artificial triplets are constructed from sampled proposal pairs.

Table 3  The performance comparison for the sgcls and sgdet task on VG150. “Model++X” is shorthand for “Model+IETrans+X”. Best overall scores in the section are highlighted in bold. Full results can be found in the appendix. (Unit: %)

5.  Discussions

In this section, we focus on the predcls task results for the Motif model3 to gain a deeper understanding of our methods. Results for RelDN can be found in appendix.

5.1  Ablation Study

Table 4 presents the components ablated from FSTA to demonstrate their contributions to the module. The results indicate that all components positively influence the improvement of F1@100. Among these, undersampling has the greatest impact on F1@100, adjusting the proportions of artificial triplets in the head, body, and tail predicate groups from 0.70, 0.14, and 0.15 to 0.33, 0.32, and 0.35 respectively. Additionally, incorporating \(\mathcal{T}_{s'po}\) effectively introduces new training combinations. The MP-sampler also plays a crucial role, further boosting R@100.

Table 4  The ablation study results for our FSTA module. For the components, “us” refers to undersampling, “+sbj” denotes adding artificial set \(\mathcal{T}_{s'po}\), and MP indicates that the MP-sampler is applied. (Unit: %)

Table 5 summarizes the ablation results with reweighting. A similar trend is observed that the components in FSTA contribute to the increase in scores for tail relation groups and the mR@100.

Table 5  The ablation study results for our FSTA module with Motif and “reweighting”. Item descriptions are identical to Table 4. (Unit: %)

5.2  Sensitivity Analysis

We then investigate the choice of percentage \(k_{s}\) in Soft Transfer. A “Naïve” setting would simply apply soft transfer to all reassigned triplets without our ranking and mapping mechanisms, where both source and target labels are assigned a value of 0.5. The results are summarized in Table 6.

Table 6  The sensitivity results for our Soft Transfer module. “\(\ast\)” indicates the setting applied in our method. (Unit: %)

As the number of entirely transferred triplets is reduced, R@100 recovers as \(k_{s}\) grows, yet mR@100 decreases. The “Naïve” setting consistently performs worse than the others in the F1@100 or even Avg@100 metrics and only be on par with the baseline IETrans (\(Avg@100=45.0\)). This highlights the significance of our devised method.

5.3  Comparison with Real Data Resampling

We compare the effects of resampling real data with our FSTA. Although both approaches augment the number of rare predicates, their motivations and methods differ. FSTA aims to steer the predicate classification layers towards understandind the inherent concept of the predicate by leveraging combinative, yet semantically plausible, artificial triplets. This also introduces new variations to the training data. In contrast, resampling merely duplicates samples from rare classes to mitigate dataset imbalance; however, it is susceptible to overfitting.

For our real data resampling implementation, we altered the training set by duplicating the image \(n\) times if it contained more than one triplet of tail group predicates. We analyzed the performance difference when applied independently and the combined for Motif+IETrans on the predcls task. The scores are listed in Table 7.

Table 7  The results of parameter choices for FSTA with Motif and “reweighting”. “\(\ast\)” indicates the setting applied in our method. (Unit: %)

Our “+FSTA” is more effective than “+resampling”, as it yields superior overall metrics for both F1 and Avg. Combining both can boost the mean recall of the tail group. Intensive resampling improved tail classes but reduced frequent class recall. We did not observe positive effects when n was larger than 4.

5.4  Parameter Choices for FSTA

To explore the quality of our \(s_{iou}\) and \(U_h\) choices, which are applied across settings, we assess the performance within the reweighting setting. Table 8 details the results.

Table 8  The results of parameter choices for FSTA with Motif and “reweighting”. “\(\ast\)” indicates the setting applied in our method. (Unit: %)

Our observations are as follows: (1) Lower values of \(U_h\) tend to result in higher tail group performance, attribute to the increased ratio of tail relations in the artificial triplets. (2) A higher \(s_{iou}\) threshold allows only the most precise feature representations, benefiting the data-sparse tail group while potentially harming the generalizability in frequent classes. Overall, we found that the parameters we selected perform reasonably well, even under such a different setting.

5.5  A Study on MP-Sampler

We examine the role of the MP-sampler. In this case, the count of artificial triplets per predicate class is invariant, but the distribution of combinations changes. We undertake a case study focusing on the mR@100 tail group, where FSTA has shown significant performance gains. Figure 6 (left) portrays the per-class recall: 12 of 17 classes either tie or improve with the MP-sampler. Next, we study the rationale behind the signal designed in the MP-sampler. It uses the scores from \(d(\cdot)\) as the sampling probability, which inversely correlates with recall. For example, \(d(\cdot)=0\) implies that the top-1 predicate prediction aligns with the ground-truth, whereas \(d(\cdot)>0\) does not. We hypothesize that sampling more object labels of higher \(d(\cdot)\) scores can make correct predictions easier for the unbised model. Thus, we analyze the changes in accumulated \(d(\cdot)\) scores between the pretrained and unbiased models, from those object classes where counts have increased. Figure 6 (right) visualizes the results computed on test data. Out of the 17 classes, 14 show non-negative total reductions scores, confirming that our designed signal is evident in the test data. We also inspect the relationship between the reduced score and class recall. The majority of classes follow a similar trend, with only four exhibiting the converse pattern (e.g., recall increases while the reduced score is negative).

Fig. 6  A case study examining the effects of the MP-sampler. (Left) The tail group mR@100 comparison between setups without and with MP-sampler. (Right) The accumulated \(d(\cdot)\) reduction contributed by objects that are sampled more frequently.

5.6  Feature Visualization

We visualize the similarity between the synthesized object features in artificial triplets and the real ones in the given classes. Figure 7 illustrates the sampled classes within the body, tail, and head groups, separated by different colors. The neighborhood identity between the real and synthetic features from the same class suggest the effectiveness of the generator.

Fig. 7  The t-SNE plots for object features in the artificial triplets. We select 10 classes from each group. Real features are plotted in dots and generated in crosses.

6.  Conclusion

In this paper, we introduce two key concepts to enhance the dataset modification approach for unbiased SGG: a novel data augmentation strategy via our FSTA module, and improved predicate reassignment efficiency through Soft Transfer. The FSTA module substantially boosts tail class recall by generating additional artificial triplets, while Soft Transfer offers a more nuanced evaluation of the reliability of individual transfers, allowing for a continuous degree of transfer and mitigating the typical decline in frequent class recall during reassignment. Experimental results confirm that integrating the complementary modules improves overall performance, surpassing the baseline IETrans.

Acknowledgments

This work was supported by the commissioned research (No. 225) by National Institute of Information and Communications Technology (NICT), JSPS/MEXT KAKENHI Grant Numbers JP23H03449 and JP22H05015.

References

[1] J. Johnson, R. Krishna, M. Stark, L.-J. Li, D.A. Shamma, M.S. Bernstein, and L. Fei-Fei, “Image retrieval using scene graphs,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp.3668-3678, 2015.
CrossRef

[2] X. Chang, P. Ren, P. Xu, Z. Li, X. Chen, and A. Hauptmann, “A comprehensive survey of scene graphs: Generation and application,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.45, no.1, pp.1-26, 2021.
CrossRef

[3] Z. Luo, W. Xie, S. Kapoor, Y. Liang, M. Cooper, J.C. Niebles, E. Adeli, and F.F. Li, “Moma: Multi-object multi-actor activity parsing,” Advances in Neural Information Processing Systems, ed. M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J.W. Vaughan, pp.17939-17955, Curran Associates, Inc., 2021.

[4] S. Aditya, Y. Yang, C. Baral, Y. Aloimonos, and C. Fermüller, “Image understanding using vision and reasoning through scene description graph,” Computer Vision and Image Understanding, vol.173, pp.33-45, 2018.
CrossRef

[5] J. Gu, S. Joty, J. Cai, H. Zhao, X. Yang, and G. Wang, “Unpaired image captioning via scene graph alignments,” Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.10323-10332, 2019.
CrossRef

[6] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y. Kalantidis, L.-J. Li, D.A. Shamma, M.S. Bernstein, and L. Fei-Fei, “Visual genome: Connecting language and vision using crowdsourced dense image annotations,” Int. J. Comput. Vision, vol.123, no.1, pp.32-73, May 2017.
CrossRef

[7] A. Desai, T.-Y. Wu, S. Tripathi, and N. Vasconcelos, “Learning of visual relations: The devil is in the tails,” Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp.15404-15413, Oct. 2021.
CrossRef

[8] B. Kang, S. Xie, M. Rohrbach, Z. Yan, A. Gordo, J. Feng, and Y. Kalantidis, “Decoupling representation and classifier for long-tailed recognition,” International Conference on Learning Representations, 2020.

[9] L. Li, L. Chen, Y. Huang, Z. Zhang, S. Zhang, and J. Xiao, “The devil is in the labels: Noisy label correction for robust scene graph generation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.18869-18878, June 2022.
CrossRef

[10] A. Zhang, Y. Yao, Q. Chen, W. Ji, Z. Liu, M. Sun, and T.-S. Chua, “Fine-grained scene graph generation with data transfer,” European conference on computer vision, pp.409-424, Springer, 2022.
CrossRef

[11] X. Lyu, L. Gao, Y. Guo, Z. Zhao, H. Huang, H.T. Shen, and J. Song, “Fine-grained predicates learning for scene graph generation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.19467-19475, June 2022.
CrossRef

[12] K. Tang, Y. Niu, J. Huang, J. Shi, and H. Zhang, “Unbiased scene graph generation from biased training,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.3716-3725, 2020.
CrossRef

[13] M.-J. Chiou, H. Ding, H. Yan, C. Wang, R. Zimmermann, and J. Feng, “Recovering the unbiased scene graphs from the biased ones,” Proceedings of the 29th ACM International Conference on Multimedia, MM ’21, New York, NY, USA, pp.1581-1590, Association for Computing Machinery, 2021.
CrossRef

[14] S. Yan, C. Shen, Z. Jin, J. Huang, R. Jiang, Y. Chen, and X.S. Hua, “Pcpl: Predicate-correlation perception learning for unbiased scene graph generation,” Proceedings of the 28th ACM International Conference on Multimedia, MM ’20, New York, NY, USA, pp.265-273, Association for Computing Machinery, 2020.
CrossRef

[15] R. Zellers, M. Yatskar, S. Thomson, and Y. Choi, “Neural motifs: Scene graph parsing with global context,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp.5831-5840, 2018.
CrossRef

[16] J. Zhang, K.J. Shih, A. Elgammal, A. Tao, and B. Catanzaro, “Graphical contrastive losses for scene graph parsing,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.11535-11543, 2019.
CrossRef

[17] C. Lu, R. Krishna, M. Bernstein, and L. Fei-Fei, “Visual relationship detection with language priors,” Computer Vision-ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, Oct. 11-14, 2016, Proceedings, Part I 14, pp.852-869, Springer, 2016.
CrossRef

[18] D. Xu, Y. Zhu, C.B. Choy, and L. Fei-Fei, “Scene graph generation by iterative message passing,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp.5410-5419, 2017.
CrossRef

[19] K. Tang, H. Zhang, B. Wu, W. Luo, and W. Liu, “Learning to compose dynamic tree structures for visual contexts,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.6619-6628, 2019.
CrossRef

[20] T. Chen, W. Yu, R. Chen, and L. Lin, “Knowledge-embedded routing network for scene graph generation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.6163-6171, 2019.
CrossRef

[21] R. Li, S. Zhang, B. Wan, and X. He, “Bipartite graph network with adaptive message passing for unbiased scene graph generation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.11109-11119, June 2021.
CrossRef

[22] J. Yu, Y. Chai, Y. Wang, Y. Hu, and Q. Wu, “Cogtree: Cognition tree loss for unbiased scene graph generation,” Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, ed. Z.H. Zhou, pp.1274-1280, International Joint Conferences on Artificial Intelligence Organization, 2021.
CrossRef

[23] I. Biederman, “Recognition-by-components: a theory of human image understanding,” Psychological review, vol.94, no.2, pp.115-147, 1987.
CrossRef

[24] P. Tokmakov, Y.-X. Wang, and M. Hebert, “Learning compositional representations for few-shot recognition,” Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.6372-6381, 2019.
CrossRef

[25] K. Kato, Y. Li, and A. Gupta, “Compositional learning for human object interaction,” Proceedings of the European Conference on Computer Vision (ECCV), pp.234-251, 2018.
CrossRef

[26] Z. Hou, X. Peng, Y. Qiao, and D. Tao, “Visual compositional learning for human-object interaction detection,” Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, Aug. 23-28, 2020, Proceedings, Part XV 16, pp.584-600, Springer, 2020.
CrossRef

[27] Z. Hou, B. Yu, Y. Qiao, X. Peng, and D. Tao, “Detecting human-object interaction via fabricated compositional learning,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.14646-14655, June 2021.
CrossRef

[28] Y. Zhong, J. Shi, J. Yang, C. Xu, and Y. Li, “Learning to generate scene graph from natural language supervision,” Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp.1823-1834, Oct. 2021.
CrossRef

[29] T. He, L. Gao, J. Song, and Y.-F. Li, “Towards open-vocabulary scene graph generation with prompt-based finetuning,” European Conference on Computer Vision, pp.56-73, Springer, 2022.
CrossRef

[30] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.

[31] Y. Xian, T. Lorenz, B. Schiele, and Z. Akata, “Feature generating networks for zero-shot learning,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp.5542-5551, 2018.
CrossRef

[32] R. Felix, B.G. Vijay Kumar, I. Reid, and G. Carneiro, “Multi-modal cycle-consistent generalized zero-shot learning,” Proceedings of the European conference on computer vision (ECCV), pp.21-37, 2018.
CrossRef

[33] B.A. Biswas and Q. Ji, “Probabilistic debiasing of scene graphs,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.10429-10438, June 2023.
CrossRef

[34] X. Han, J. Yang, H. Hu, L. Zhang, J. Gao, and P. Zhang, “Image scene graph generation (sgg) benchmark,” arXiv preprint arXiv:2107.12604, 2021.

[35] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp.770-778, 2016.
CrossRef

[36] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, vol.28, 2015.

[37] M. Arjovsky and L. Bottou, “Towards principled methods for training generative adversarial networks,” arXiv preprint arXiv:1701.04862, 2017.

[38] A. Radford, J.W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, G. Krueger, and I. Sutskever, “Learning transferable visual models from natural language supervision,” CoRR, vol.abs/2103.00020, 2021.

Appendix A: Full Sgcls and Sgdet Tasks Results

The full results for the “sgcls” and “sgdet” tasks are listed in Table A\(\cdot\)5 (sgcls) and A\(\cdot\)6 (sgdet).

Appendix B: Qualitative Results

We visualize the results of predicate prediction in Fig. A\(\cdot\)1.

Fig. A・1  Qualitative results of our method for the predcls task under the Motif+rwt setting: (Left) Images with bounding boxes, (Middle) Ground-truth scene graphs, and (Right) Predicted results. Isolated nodes have been omitted from the visualized scene graphs. The relations in red indicate discrepancies with the ground truth.

Appendix C: Results for Ablation Study (RelDN)

Table A\(\cdot\)1 and Table A\(\cdot\)2 summarize the ablation results for FSTA module under RelDN and RelDN with reweighting, respectively.

Table A・1  The ablation study results for our FSTA module with RelDN. Item descriptions are identical to Table 4. (Unit: %)

Table A・2  The ablation study results for our FSTA module with RelDN and “reweighting”. Item descriptions are identical to Table 4. (Unit: %)

Appendix D: Results for Sensitivity Analysis (RelDN)

Table A\(\cdot\)3 lists the Soft Transfer sensitivity results of RelDN for the predcls task. \(k_{s}=0.7\) actually achieves a better score on both R@100 and mR@100 over the baseline method (an improvement of 39.9 to 40.8 for R@100 and 32.3 to 32.5 for mR@100). Nevertheless, modifying \(Q(\cdot)=1-Q'(\cdot)\) to \(Q(\cdot)=Q'(\cdot)\) leads to stronger overall performance, due to the larger space for the R@100 score recovery. Again, the “Naïve” case is inferior to the applied settings.

Table A・3  The sensitivity results with RelDN for our Soft Transfer module. “\(\ast\)” indicates the setting applied in our method. “\(\diamond\)” stands for a modified \(Q(\cdot)\). (Unit: %)

Appendix E: Results for FSTA Parameter Choices (RelDN)

Table A\(\cdot\)4 describes the results of FSTA parameter study for RelDN.

Table A・4  The results of parameter choices for FSTA with RelDN and “reweighting”. “\(\ast\)” indicates the setting applied in our method. (Unit: %)

Appendix F: Randomness of FSTA

The resource of randomness: Including the undersampling step and the generator pretraining. We selected a fixed checkpoint for the generator based on the classification accuracies observed in the validation data.

The reproducibility of randomness: In the SGG model training, we followed the open-source SGG model implementation1 to set the seed for the libraries and switch to deterministic mode for the cudnn library.

The impact of randomness: We measured the standard deviation of R@100 and mR@100 under “Motif++FSTA+rwt” in the predcls task, using five different runs. The values are 0.23 for R@100 and 0.31 for mR@100. Note that these include randomness from both the Motif model and the FSTA module.

Table A・5  The full performance comparison for the sgcls task on VG150. Scores for models listed in the first section are cited from their original papers, while models in subsequent sections use our implementation. “Model++X” is shorthand for “Model+IETrans+X”. The best overall scores within each section are highlighted in bold. (Unit: %)

Table A・6  The full performance comparison for the sgdet task on VG150. Scores for models listed in the first section are cited from their original papers, while models in subsequent sections use our implementation. “Model++X” is shorthand for “Model+IETrans+X”. The best overall scores within each section are highlighted in bold. (Unit: %)

Appendix G: Object Generator

We exploit a conditional-GAN based model to synthesize object' features, due to its lightweight and low additional computational cost. In the pre-processing step, we collect the real features from model predictions on training data (See Fig. 4 in the manuscript). The adversarial loss function for the GAN model consist of three parts: \(\mathcal{L}_{wgangp}\), \(\mathcal{L}_{cls}\), and \(\mathcal{L}_{recon}\).

\(\mathcal{L}_{wgangp}\) is a standard WGAN loss with gradient penalty [37] as Eq. (A\(\cdot\) 1).

\[\begin{align} & \mathcal{L}_{wgangp} = \mathbb{E}_{\mathbf{x} \sim real}[D(\mathbf{x}, \mathbf{s}_{c})] - \mathbb{E}_{\tilde{\mathbf{x}} \sim gen}[D(\tilde{\mathbf{x}}, \mathbf{s}_{c})] \nonumber \\ & \hphantom{\mathcal{L}_{wgangp} =} - \lambda \mathbb{E}[(||\nabla_{\hat{\mathbf{x}}}D(\hat{\mathbf{x}}, \mathbf{s}_{c})||_{2} -1)^{2}] \tag{A$\cdot $1} \end{align}\]

where \(\mathbf{x} \in \mathbb{R}^{d}\) is the feature sampled from real data, \(\tilde{\mathbf{x}} = G(\mathbf{z}, \mathbf{s}_{c}) \in \mathbb{R}^{d}\) is the synthesized feature from generator \(G\). \(d\) is the size of object feature. \(\hat{\mathbf{x}} = \alpha \mathbf{x} + (1 - \alpha) \tilde{\mathbf{x}}\) is an interpolated feature with \(\alpha\) sampled from a uniform distribution. \(\mathbf{z}\) is an initial vector sampled from normal distribution, and \(\mathbf{s}_{c}\) is a condition vector represents the object class. We collect \(\mathbf{s}_{c}\) from the pre-trained CLIP [38] text encoder. We use the basic template “a photo of a [OBJECT NAME].” as the input prompt to text encoder, then the output vector as the class representation.

\(\mathcal{L}_{cls}\) is a regularization loss for the generator \(G\). It utilizes a softmax classifier pre-trained on real data to encourage the generator to output features with enhanced discriminability. That is, the synthetic features can be better classified. Equation (A\(\cdot\) 2) describes its loss function.

\[\begin{align} \mathcal{L}_{cls} = - \mathbb{E}_{\tilde{\mathbf{x}} \sim gen}[\mathrm{log} P(y|\tilde{x}; \theta_{cls})] \tag{A$\cdot $2} \end{align}\]

where \(\theta_{cls}\) is the weights of the softmax classifier. \(y\) is the corresponding class label. During the adversarial training, the pre-trained classifier is frozen.

\(\mathcal{L}_{recon}\) is another regularization term for the class consistency between generator output and its condition input. A reconstructor \(R(\cdot)\) is pre-trained on real data to infer the class condition vector from the feature. Equation (A\(\cdot\) 3) describes its loss function.

\[\begin{align} \mathcal{L}_{recon} = \mathbb{E}_{\tilde{\mathbf{x}} \sim gen}[\lVert R(\tilde{\mathbf{x}}) - \mathbf{s}_{c} \rVert_{2}] \tag{A$\cdot $3} \end{align}\]

the reconstructor is also frozen in the adversarial training.

The overall loss function is as below and identical to Eq. (4) in the main paper.

\[\begin{align} \min_{G}\max_{D} \mathcal{L}_{wgangp} + \beta \mathcal{L}_{cls} + \gamma \mathcal{L}_{recon} \tag{A$\cdot $4} \end{align}\]

We list the model architecture for training the object generator in Table A\(\cdot\)7.

Table A・7  The model architecture.

Appendix H: Hyperparameter Details

We list the parameter choices for training SGG models and the generator model in Table A\(\cdot\)8.

Table A・8  The parameter choices for training Motif-based SGG models (section 1), RelDN-based SGG models (section 2), and the genertor model (section 3).

Footnotes

1. https://github.com/microsoft/scene_graph_benchmark

2. We follow the implementations in 1 to compute \(s_{iou}\).

3. Unless specified otherwise, this means “Model+IETrans+X”.

Authors

KuanChao CHU
  The University of Tokyo

he is currently pursuing the doctoral degree with the Nakayama Laboratory, Graduate School of Information Science and Technology, The University of Tokyo. His research interests include novel object detection, data augmentation, scene graph detection, and deep learning.

Satoshi YAMAZAKI
  NEC Corporation

the master’s and Ph.D. degrees in physics from Tohoku University, Japan, in 2011 and 2014, respectively. He is researcher in NEC Corporation since April 2014. His research interests include object tracking, person re-identification, and scene graph generation, and deep learning.

Hideki NAKAYAMA
  The University of Tokyo

received the master’s and Ph.D. degrees in information science from the University of Tokyo, Japan, in 2008 and 2011, respectively. From 2012 to 2018, he was an Assistant Professor at the Graduate School of Information Science and Technology, The University of Tokyo, where he has been an Associate Professor, since April 2018. He is also a Faculty Member with the International Research Center for Neurointelligence (IRCN) and a Visiting Researcher with the National Institute of Advanced Industrial Science and Technology (AIST). His research interests include generic image recognition, natural language processing, and deep learning.

Keyword