研究者業績

計良 宥志

ケラ ヒロシ  (Hiroshi Kera)

基本情報

所属
千葉大学 大学院情報学研究院 助教
学位
博士(2020年3月 東京大学)

研究者番号
00887705
ORCID ID
 https://orcid.org/0000-0002-9830-0436
J-GLOBAL ID
202001000989405698
researchmap会員ID
R000013274

論文

 30
  • Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
    Neural Information Processing Systems 2024年12月  査読有り
  • Hiroshi Kera, Yuki Ishihara, Yuta Kambe, Tristan Vaccon, Kazuhiro Yokoyama
    Neural Information Processing Systems abs/2311.12904 2024年12月  査読有り筆頭著者責任著者
  • Hiroshi Kera, Yuki Ishihara, Tristan Vaccon, Kazuhiro Yokoyama
    International Symposium on Symbolic Computation in Software Science (SCSS 2024), Work in Progress Workshop 51-56 2024年8月  査読有り筆頭著者責任著者
  • Hiroshi Kera, Yoshihiko Hasegawa
    Journal of Computational Algebra 100022-100022 2024年8月  査読有り筆頭著者責任著者
  • Ayumu Ueyama, Kazuhiko Kawamoto, Hiroshi Kera
    International Conference on Machine Learning Workshop on Machine Learning for Earth System Modeling (ICML ML4ESM Workshop) 2024年7月  査読有り最終著者責任著者
    Weather forecasting is essential for various human activities. Recent data-driven models have outperformed numerical weather prediction by utilizing deep learning in forecasting performance. However, challenges remain in efficiently handling multiple meteorological variables. This study proposes a new variable aggregation scheme and an efficient learning framework for that challenge. Experiments show that VarteX outperforms the conventional model in forecast performance, requiring significantly fewer parameters and resources. The effectiveness of learning through multiple aggregations and regional split training is demonstrated, enabling more efficient and accurate deep learning-based weather forecasting.
  • Yusuke Marumo, Kazuhiko Kawamoto, Hiroshi Kera
    IEEE/CFV Conference on Computer Vision and Pattern Recognition Workshop on Computer Vision for Animals (CVPR Workshop on CV4A) 2024年6月  査読有り最終著者責任著者
  • Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
    International Conference on Learning Representations 2024年5月  査読有り
    It is not fully understood why adversarial examples can deceive neural networks and transfer between different networks. To elucidate this, several studies have hypothesized that adversarial perturbations, while appearing as noises, contain class features. This is supported by empirical evidence showing that networks trained on mislabeled adversarial examples can still generalize well to correctly labeled test samples. However, a theoretical understanding of how perturbations include class features and contribute to generalization is limited. In this study, we provide a theoretical framework for understanding learning from perturbations using a one-hidden-layer network trained on mutually orthogonal samples. Our results highlight that various adversarial perturbations, even perturbations of a few pixels, contain sufficient class features for generalization. Moreover, we reveal that the decision boundary when learning from perturbations matches that from standard samples except for specific regions under mild conditions. The code is available at https://github.com/s-kumano/learning-from-adversarial-perturbations.
  • Kosuke Sumiyasu, Kazuhiko Kawamoto, Hiroshi Kera
    IEEE/CFV Conference on Computer Vision and Pattern Recognition (CVPR) 6017-6026 2024年1月8日  査読有り最終著者責任著者
    To better understand the behavior of image classifiers, it is useful to visualize the contribution of individual pixels to the model prediction. In this study, we propose a method, MoXI ($\textbf{Mo}$del e$\textbf{X}$planation by $\textbf{I}$nteractions), that efficiently and accurately identifies a group of pixels with high prediction confidence. The proposed method employs game-theoretic concepts, Shapley values and interactions, taking into account the effects of individual pixels and the cooperative influence of pixels on model confidence. Theoretical analysis and experiments demonstrate that our method better identifies the pixels that are highly contributing to the model outputs than widely-used visualization by Grad-CAM, Attention rollout, and Shapley value. While prior studies have suffered from the exponential computational cost in the computation of Shapley value and interactions, we show that this can be reduced to quadratic cost for our task. The code is available at https://github.com/KosukeSumiyasu/MoXI.
  • Nariki Tanaka, Hiroshi Kera, Kazuhiko Kawamoto
    Comput. Vis. Image Underst. 240 103936-103936 2024年  査読有り
    Using Fourier analysis, we explore the robustness and vulnerability of graph convolutional neural networks (GCNs) for skeleton-based action recognition. We adopt a joint Fourier transform (JFT), a combination of the graph Fourier transform (GFT) and the discrete Fourier transform (DFT), to examine the robustness of adversarially-trained GCNs against adversarial attacks and common corruptions. Experimental results with the NTU RGB+D dataset reveal that adversarial training does not introduce a robustness trade-off between adversarial attacks and low-frequency perturbations, which typically occurs during image classification based on convolutional neural networks. This finding indicates that adversarial training is a practical approach to enhancing robustness against adversarial attacks and common corruptions in skeleton-based action recognition. Furthermore, we find that the Fourier approach cannot explain vulnerability against skeletal part occlusion corruption, which highlights its limitations. These findings extend our understanding of the robustness of GCNs, potentially guiding the development of more robust learning methods for skeleton-based action recognition.
  • Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
    Neural Information Processing Systems 2023年12月  査読有り
    Spotlight
  • Nan Wu, Hiroshi Kera, Kazuhiko Kawamoto
    Communications in Computer and Information Science 18-28 2023年10月30日  査読有り
  • Tomoyasu Nanaumi, Kazuhiko Kawamoto, Hiroshi Kera
    International Conference on Computer Vision Workshop on Uncertainty in Computer Vision (ICCV UnCV Workshop) 2023年10月  査読有り最終著者責任著者
  • Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
    Pattern Recognit. Lett. 172 259-265 2023年8月  査読有り
  • Nan Wu, Hiroshi Kera, Kazuhiko Kawamoto
    Applied Intelligence 53(20) 24142-24156 2023年7月17日  査読有り
    Zero-shot action recognition, which recognizes actions in videos without having received any training examples, is gaining wide attention considering it can save labor costs and training time. Nevertheless, the performance of zero-shot learning is still unsatisfactory, which limits its practical application. To solve this problem, this study proposes a framework to improve zero-shot action recognition using human instructions with text descriptions. The proposed framework manually describes video contents, which incurs some labor costs; in many situations, the labor costs are worth it. We manually annotate text features for each action, which can be a word, phrase, or sentence. Then by computing the matching degrees between the video and all text features, we can predict the class of the video. Furthermore, the proposed model can also be combined with other models to improve its accuracy. In addition, our model can be continuously optimized to improve the accuracy by repeating human instructions. The results with UCF101 and HMDB51 showed that our model achieved the best accuracy and improved the accuracies of other models.
  • Chun Yang Tan, Kazuhiko Kawamoto, Hiroshi Kera
    European Conference on Computer Vision Workshop on Adversarial Robustness in the Real World (ECCV AROW Workshop) 2023年5月15日  査読有り最終著者責任著者
    In recent years, there has been growing concern over the vulnerability of convolutional neural networks (CNNs) to image perturbations. However, achieving general robustness against different types of perturbations remains challenging, in which enhancing robustness to some perturbations (e.g., adversarial perturbations) may degrade others (e.g., common corruptions). In this paper, we demonstrate that adversarial training with an emphasis on phase components significantly improves model performance on clean, adversarial, and common corruption accuracies. We propose a frequency-based data augmentation method, Adversarial Amplitude Swap, that swaps the amplitude spectrum between clean and adversarial images to generate two novel training images: adversarial amplitude and adversarial phase images. These images act as substitutes for adversarial images and can be implemented in various adversarial training setups. Through extensive experiments, we demonstrate that our method enables the CNNs to gain general robustness against different types of perturbations and results in a uniform performance against all types of common corruptions.
  • Elias Samuel Wirth, Hiroshi Kera, Sebastian Pokutta
    International Conference on Learning Representations 2023年5月  査読有り
    The vanishing ideal of a set of points $X = \{\mathbf{x}_1, \ldots, \mathbf{x}_m\}\subseteq \mathbb{R}^n$ is the set of polynomials that evaluate to $0$ over all points $\mathbf{x} \in X$ and admits an efficient representation by a finite subset of generators. In practice, to accommodate noise in the data, algorithms that construct generators of the approximate vanishing ideal are widely studied but their computational complexities remain expensive. In this paper, we scale up the oracle approximate vanishing ideal algorithm (OAVI), the only generator-constructing algorithm with known learning guarantees. We prove that the computational complexity of OAVI is not superlinear, as previously claimed, but linear in the number of samples $m$. In addition, we propose two modifications that accelerate OAVI's training time: Our analysis reveals that replacing the pairwise conditional gradients algorithm, one of the solvers used in OAVI, with the faster blended pairwise conditional gradients algorithm leads to an exponential speed-up in the number of features $n$. Finally, using a new inverse Hessian boosting approach, intermediate convex optimization problems can be solved almost instantly, improving OAVI's training time by multiple orders of magnitude in a variety of numerical experiments.
  • Takeshi Haga, Hiroshi Kera, Kazuhiko Kawamoto
    Sensors 23(5) 2515-2515 2023年3月  査読有り
    In this paper, we propose a sequential variational autoencoder for video disentanglement, which is a representation learning method that can be used to separately extract static and dynamic features from videos. Building sequential variational autoencoders with a two-stream architecture induces inductive bias for video disentanglement. However, our preliminary experiment demonstrated that the two-stream architecture is insufficient for video disentanglement because static features frequently contain dynamic features. Additionally, we found that dynamic features are not discriminative in the latent space. To address these problems, we introduced an adversarial classifier using supervised learning into the two-stream architecture. The strong inductive bias through supervision separates dynamic features from static features and yields discriminative representations of the dynamic features. Through a comparison with other sequential variational autoencoders, we qualitatively and quantitatively demonstrate the effectiveness of the proposed method on the Sprites and MUG datasets.
  • Hiroshi Kera
    ISSAC '22: International Symposium on Symbolic and Algebraic Computation(ISSAC) 225-234 2022年7月  筆頭著者最終著者責任著者
  • Soichiro Kumano, Hiroshi Kera, Toshihiko Yamasaki
    2022年5月29日  査読有り
    Adversarial attacks have only focused on changing the predictions of the classifier, but their danger greatly depends on how the class is mistaken. For example, when an automatic driving system mistakes a Persian cat for a Siamese cat, it is hardly a problem. However, if it mistakes a cat for a 120km/h minimum speed sign, serious problems can arise. As a stepping stone to more threatening adversarial attacks, we consider the superclass adversarial attack, which causes misclassification of not only fine classes, but also superclasses. We conducted the first comprehensive analysis of superclass adversarial attacks (an existing and 19 new methods) in terms of accuracy, speed, and stability, and identified several strategies to achieve better performance. Although this study is aimed at superclass misclassification, the findings can be applied to other problem settings involving multiple classes, such as top-k and multi-label classification attacks.
  • Takaaki Azakami, Hiroshi Kera, Kazuhiko Kawamoto
    International Conference on Systems, Man, and Cybernetics (SMC 2022) 2022年5月20日  査読有り
    We propose an evolutionary computation method for an adversarial attack on the length and thickness of parts of legged robots by deep reinforcement learning. This attack changes the robot body shape and interferes with walking-we call the attacked body as adversarial body shape. The evolutionary computation method searches adversarial body shape by minimizing the expected cumulative reward earned through walking simulation. To evaluate the effectiveness of the proposed method, we perform experiments with three-legged robots, Walker2d, Ant-v2, and Humanoid-v2 in OpenAI Gym. The experimental results reveal that Walker2d and Ant-v2 are more vulnerable to the attack on the length than the thickness of the body parts, whereas Humanoid-v2 is vulnerable to the attack on both of the length and thickness. We further identify that the adversarial body shapes break left-right symmetry or shift the center of gravity of the legged robots. Finding adversarial body shape can be used to proactively diagnose the vulnerability of legged robot walking.
  • Takuto Otomo, Hiroshi Kera, Kazuhiko Kawamoto
    International Conference on Systems, Man, and Cybernetics (SMC 2022) 2022年5月20日  査読有り
    We address adversarial attacks on the actuators at the joints of legged robots trained by deep reinforcement learning. The vulnerability to the joint attacks can significantly impact the safety and robustness of legged robots. In this study, we demonstrate that the adversarial perturbations to the torque control signals of the actuators can significantly reduce the rewards and cause walking instability in robots. To find the adversarial torque perturbations, we develop black-box adversarial attacks, where, the adversary cannot access the neural networks trained by deep reinforcement learning. The black box attack can be applied to legged robots regardless of the architecture and algorithms of deep reinforcement learning. We employ three search methods for the black-box adversarial attacks: random search, differential evolution, and numerical gradient descent methods. In experiments with the quadruped robot Ant-v2 and the bipedal robot Humanoid-v2, in OpenAI Gym environments, we find that differential evolution can efficiently find the strongest torque perturbations among the three methods. In addition, we realize that the quadruped robot Ant-v2 is vulnerable to the adversarial perturbations, whereas the bipedal robot Humanoid-v2 is robust to the perturbations. Consequently, the joint attacks can be used for proactive diagnosis of robot walking instability.
  • Nariki Tanaka, Hiroshi Kera, Kazuhiko Kawamoto
    Proceedings of the AAAI Conference on Artificial Intelligence, 2022 2335-2343 2022年2月22日  査読有り
    Skeleton-based action recognition models have recently been shown to be vulnerable to adversarial attacks. Compared to adversarial attacks on images, perturbations to skeletons are typically bounded to a lower dimension of approximately 100 per frame. This lower-dimensional setting makes it more difficult to generate imperceptible perturbations. Existing attacks resolve this by exploiting the temporal structure of the skeleton motion so that the perturbation dimension increases to thousands. In this paper, we show that adversarial attacks can be performed on skeleton-based action recognition models, even in a significantly low-dimensional setting without any temporal manipulation. Specifically, we restrict the perturbations to the lengths of the skeleton's bones, which allows an adversary to manipulate only approximately 30 effective dimensions. We conducted experiments on the NTU RGB+D and HDM05 datasets and demonstrate that the proposed attack successfully deceived models with sometimes greater than 90\% success rate by small perturbations. Furthermore, we discovered an interesting phenomenon: in our low-dimensional setting, the adversarial training with the bone length attack shares a similar property with data augmentation, and it not only improves the adversarial robustness but also improves the classification accuracy on the original original data. This is an interesting counterexample of the trade-off between adversarial robustness and clean accuracy, which has been widely observed in studies on adversarial training in the high-dimensional regime.
  • Kevin Richard G. Operiano, Wanchalerm Pora, Hitoshi Iba, Hiroshi Kera
    IEEE Access 9 164379-164393 2021年  査読有り最終著者
  • Hiroshi Kera, Yoshihiko Hasegawa
    The Thirty-Fourth AAAI Conference on Artificial Intelligence(AAAI) 4428-4435 2020年2月  査読有り筆頭著者
  • Hiroshi Kera, Yoshihiko Hasegawa
    IEEE Access 7 178961-178976 2019年1月25日  査読有り筆頭著者
    Approximate vanishing ideal is a concept from computer algebra that studies the algebraic varieties behind perturbed data points. To capture the nonlinear structure of perturbed points, the introduction of approximation to exact vanishing ideals plays a critical role. However, such an approximation also gives rise to a theoretical problem---the spurious vanishing problem---in the basis construction of approximate vanishing ideals; namely, obtained basis polynomials can be approximately vanishing simply because of the small coefficients. In this paper, we propose a first general method that enables various basis construction algorithms to overcome the spurious vanishing problem. In particular, we integrate coefficient normalization with polynomial-based basis constructions, which do not need the proper ordering of monomials to process for basis constructions. We further propose a method that takes advantage of the iterative nature of basis construction so that computationally costly operations for coefficient normalization can be circumvented. Moreover, a coefficient truncation method is proposed for further accelerations. From the experiments, it can be shown that the proposed method overcomes the spurious vanishing problem, resulting in shorter feature vectors while sustaining comparable or even lower classification error.
  • Hiroshi Kera, Yoshihiko Hasegawa
    Proceedings of the AAAI Conference on Artificial Intelligence 32 3399-3406 2018年1月29日  査読有り筆頭著者
    The vanishing ideal is a set of polynomials that takes zero value on the given data points. Originally proposed in computer algebra, the vanishing ideal has been recently exploited for extracting the nonlinear structures of data in many applications. To avoid overfitting to noisy data, the polynomials are often designed to approximately rather than exactly equal zero on the designated data. Although such approximations empirically demonstrate high performance, the sound algebraic structure of the vanishing ideal is lost. The present paper proposes a vanishing ideal that is tolerant to noisy data and also pursued to have a better algebraic structure. As a new problem, we simultaneously find a set of polynomials and data points for which the polynomials approximately vanish on the input data points, and almost exactly vanish on the discovered data points. In experimental classification tests, our method discovered much fewer and lower-degree polynomials than an existing state-of-the-art method. Consequently, our method accelerated the runtime of the classification tasks without degrading the classification accuracy.
  • Yifei Huang, Minjie Cai, Hiroshi Kera, Ryo Yonetani, Keita Higuchi, Yoichi Sato
    2017 IEEE International Conference on Computer Vision Workshops (ICCVW) 2313-2321 2017年10月  査読有り
  • Hiroshi Kera, Hitoshi Iba
    2016 IEEE Congress on Evolutionary Computation (CEC) 5018-5025 2016年7月  査読有り筆頭著者
  • Hiroshi Kera, Yoshihiko Hasegawa
    Nonlinear Dynamics 85(1) 675-692 2016年7月  査読有り筆頭著者
  • Hiroshi Kera, Ryo Yonetani, Keita Higuchi, Yoichi Sato
    2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 361-369 2016年6月  査読有り筆頭著者

MISC

 5
  • Yusuke Marumo, Kazuhiko Kawamoto, Hiroshi Kera
    CoRR abs/2403.08227 2024年  最終著者責任著者
    Not identical but similar objects are everywhere in the world. Examples include four-legged animals such as dogs and cats, cars of different models, akin flowers in various colors, and countless others. In this study, we address a novel task of matching such non-identical objects. We propose a simple weighting scheme of descriptors that enhances various sparse image matching methods, which were originally designed for matching identical objects captured from different perspectives, and achieve semantically robust matching. The experiments show successful matching between non-identical objects in various cases including domain shift. Further, we present a first evaluation of the robustness of the image matching methods under common corruptions, which is a sort of domain shift, and the proposed method improves the matching in this case as well.
  • Ryosuke Masuya, Yuichi Ike, Hiroshi Kera
    2022年10月27日  最終著者責任著者
    Vanishing component analysis (VCA) computes approximate generators of vanishing ideals of samples, which are further used for extracting nonlinear features of the samples. Recent studies have shown that normalization of approximate generators plays an important role and different normalization leads to generators of different properties. In this paper, inspired by recent self-supervised frameworks, we propose a contrastive normalization method for VCA, where we impose the generators to vanish on the target samples and to be normalized on the transformed samples. We theoretically show that a contrastive normalization enhances the discriminative power of VCA, and provide the algebraic interpretation of VCA under our normalization. Numerical experiments demonstrate the effectiveness of our method. This is the first study to tailor the normalization of approximate generators of vanishing ideals to obtain discriminative features.
  • Kosuke Sumiyasu, Kazuhiko Kawamoto, Hiroshi Kera
    2022年10月7日  最終著者責任著者
    This paper analyzes various types of image misclassification from a game-theoretic view. Particularly, we consider the misclassification of clean, adversarial, and corrupted images and characterize it through the distribution of multi-order interactions. We discover that the distribution of multi-order interactions varies across the types of misclassification. For example, misclassified adversarial images have a higher strength of high-order interactions than correctly classified clean images, which indicates that adversarial perturbations create spurious features that arise from complex cooperation between pixels. By contrast, misclassified corrupted images have a lower strength of low-order interactions than correctly classified clean images, which indicates that corruptions break the local cooperation between pixels. We also provide the first analysis of Vision Transformers using interactions. We found that Vision Transformers show a different tendency in the distribution of interactions from that in CNNs, and this implies that they exploit the features that CNNs do not use for the prediction. Our study demonstrates that the recent game-theoretic analysis of deep learning models can be broadened to analyze various malfunctions of deep learning models including Vision Transformers by using the distribution, order, and sign of interactions.
  • Chun Yang Tan, Kazuhiko Kawamoto, Hiroshi Kera
    2022年3月14日  最終著者責任著者
    The vulnerability of convolutional neural networks (CNNs) to image perturbations such as common corruptions and adversarial perturbations has recently been investigated from the perspective of frequency. In this study, we investigate the effect of the amplitude and phase spectra of adversarial images on the robustness of CNN classifiers. Extensive experiments revealed that the images generated by combining the amplitude spectrum of adversarial images and the phase spectrum of clean images accommodates moderate and general perturbations, and training with these images equips a CNN classifier with more general robustness, performing well under both common corruptions and adversarial perturbations. We also found that two types of overfitting (catastrophic overfitting and robust overfitting) can be circumvented by the aforementioned spectrum recombination. We believe that these results contribute to the understanding and the training of truly robust classifiers.
  • Wataru Okamoto, Hiroshi Kera, Kazuhiko Kawamoto
    2021年11月19日  
    This study is aimed at addressing the problem of fault tolerance of quadruped robots to actuator failure, which is critical for robots operating in remote or extreme environments. In particular, an adaptive curriculum reinforcement learning algorithm with dynamics randomization (ACDR) is established. The ACDR algorithm can adaptively train a quadruped robot in random actuator failure conditions and formulate a single robust policy for fault-tolerant robot control. It is noted that the hard2easy curriculum is more effective than the easy2hard curriculum for quadruped robot locomotion. The ACDR algorithm can be used to build a robot system that does not require additional modules for detecting actuator failures and switching policies. Experimental results show that the ACDR algorithm outperforms conventional algorithms in terms of the average reward and walking distance.

主要な講演・口頭発表等

 10

主要な共同研究・競争的資金等の研究課題

 8
  • 科学技術振興機構 ムーンショット型研究開発事業・目標8 (コア研究) 2023年12月 - 2027年3月
    小槻峻司(代表), 大塚敏之, 小蔵正輝, 松岡大祐, 計良宥志, 徳田慶太, 薄 良彦, 井元佑介, 岡﨑淳史, 金丸佳矢, 安永数明, 濱田 篤, 平賀優介, 増永浩彦, 重本達哉, 近藤卓也, 福重さと子, 嘉村雄司, 堀 智晴, 山田真史, 山田進二, 風間 聡, 立花幸司, 舩冨卓也, 久保尋之
  • 科学技術振興機構 戦略的な研究開発の推進 戦略的創造研究推進事業 ACT-X 2023年 - 2025年
    計良 宥志
  • 科学技術振興機構 戦略的な研究開発の推進 戦略的創造研究推進事業 ACT-X 2020年10月 - 2022年
    計良 宥志