Updated on 2022/05/20

写真a

 
OHYA, Jun
 
Affiliation
Faculty of Science and Engineering, School of Creative Science and Engineering
Job title
Professor

Concurrent Post

  • Faculty of Science and Engineering   School of Fundamental Science and Engineering

  • Faculty of Science and Engineering   Graduate School of Creative Science and Engineering

  • Affiliated organization   Global Education Center

Research Institute

  • 2020
    -
    2022

    理工学術院総合研究所   兼任研究員

  • 2020
    -
    2022

    国際情報通信研究センター   兼任研究員

Education

  •  
    -
    1979

    University of Tokyo   Graduate School, Division of Engineering   Department of Precision Machinery Engineering  

  •  
    -
    1977

    University of Tokyo   Faculty of Engineering   Department of Precision Machinery Engineering  

Degree

  • University of Tokyo   Ph.D. (Dr. Engineering)-

Research Experience

  • 2014
    -
    Now

    Waseda Uniniversity, Department of Modern Mechanical Engineering   Professsor

  • 2002
    -
    2013

    present Waseda University, Graduate School of Global Information and Telecommunication Studies, Professor

  • 2005.07
    -
    2005.09

    Visiting professor, the University of Karlsruhe, Germany

  • 2000
    -
    2002

    Waseda University, Global Information and Telecommunication Institute, Professor

  • 1996
    -
    2000

    ATR Media Integration & Communication Research Laboratories, Department head

  • 1992
    -
    1996

    ATR Communication Systems Research Laboratories, Supervising researcher

  • 1989
    -
    1992

    NTT Human Interface Laboratories, Supervising researcher

  • 1988
    -
    1989

    University of Maryland, U.S.A.,Visiting research associate

  • 1979
    -
    1988

    NTT Electrical Communications Laboratories,Researcher

▼display all

Professional Memberships

  •  
     
     

    The Institute of Image Electronics Engineers of Japan

  •  
     
     

    The Virtual Reality Society of Japan

  •  
     
     

    Information Processing Society of Japan

  •  
     
     

    The Institute of Electronics, Information and Communication Engineers

  •  
     
     

    Institute of Electronics, Information and Communication Engineers

 

Research Areas

  • Intelligent informatics

Research Interests

  • Computer vision,Computer graphics,Virtual reality,Computer Vision,Virtual Reality,Multimedia Tele-communication

Papers

  • Estimating the 3D Cut Position of Pork Frontal Legs in RGBD Images by a Deep Learning Based Method for Achieving a Robot That Cuts Pork Legs Autonomously

    Ryo Onuki, Taiki Suzuki, Jeonghwang Hayashi, Chanjin Seo, Jun Ohya, Takaaki Ohkawauchi

    The 7th IIEEJ International Conference on Image Electronics and Visual Computing   ( 3A-4 ) 1 - 4  2021.09  [Refereed]

  • Early Detection of Objects on the Road Using V2V and Multiple Image Streams for Supervised Deep-Learning Based Autonomous Driving

    Zelin Zhang, Jun Ohya

    The 7th IIEEJ International Conference on Image Electronics and Visual Computing   ( 3A-3 ) 1 - 4  2021.09  [Refereed]

  • Interpreting Evaluation Criteria of Running Foot Strike Classifier Using Kinematic Model

    Chanjin Seo, Yuta Goto, Koji Tagami, Hiroyuki Ogata, Kazuyuki Kanosue, Jun Ohya

    The 7th IIEEJ International Conference on Image Electronics and Visual Computing   ( 1A-4 ) 1 - 4  2021.09  [Refereed]

  • Movement Control with Vehicle-to-Vehicle Communication by Using End-to-End Deep Learning for Cooperative Autonomous Driving

    Zelin, Zhang, Jun OHYA

    10th International Conference on Pattern Recognition Applications and Methods     377 - 385  2021.02  [Refereed]

  • Quantitative Method for Evaluating the Coordination between Sprinting Motions using Joint Coordinates Obtained from the Videos and Cross-correlations

    Masato Sabanai, Chanjin Seo, Hiroyuki Ogata, Jun Ohya

    10th International Conference on Pattern Recognition Applications and Methods     531 - 539  2021.02  [Refereed]

  • Data Collection Through Translation Network Based on End-to-End Deep Learning for Autonomous Driving

    Zelin, Zhang, Jun OHYA

    IS&T International Symposium on Electronic Imaging. (EI 2021)     115 - 121  2021.01  [Refereed]

  • Extracting and Interpreting Unknown Factors with Classifier for Foot Strike Types in Running

    Chanjin Seo, Masato Sabanai, Yuta Goto, Koji Tagami, Hiroyuki Ogata, Kazuyuki Kanosue, Jun Ohya

    25th International Conference on Pattern Recognition (ICPR)     3217 - 3224  2021.01  [Refereed]

  • Estimating the Orientation of 3D-Structured Switches Using the RGB-D Camera Attached to a Disaster Response Robot

    Takuya Kanda, Li Qi, Jun Ohya, Atsuo Takanishi, Hiroyuki Ogata, Kenji Hshimoto

    International Journal of Mechanical Engineering and Robotics Research   9 ( 10 ) 1432 - 1438  2020.10  [Refereed]

  • Developing Thermal Endoscope for Endoscopic Photothermal Therapy for Peritoneal Dissemination

    Mutsuki Ohara, Sohta Sanpei, Chanjin Seo, Jun Ohya, Ken Masamune, Hiroshi Nagahashi, Yuji Morimoto, Manabu Harada

    2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)     3040 - 3047  2020.10  [Refereed]

  • Automatic Estimation of the Position and Orientation of Stairs to Be Reached and Climbed by a Disaster Response Robot by Analyzing 2D Image and 3D Point Cloud

    Kazuya Miyakawa, Takuya Kanda, Jun Ohya, Hiroyuki Ogata, Kenji Hashimoto, Atsuo Takanishi

    International Journal of Mechanical Engineering and Robotics Research   9 ( 9 ) 1312 - 1321  2020.09  [Refereed]

  • Classification of Aortic Stenosis Using ECG by Deep Learning and its Analysis Using Grad-CAM

    Erika Hata, Chanjin Seo, Masafumi Nakayama, Kiyotaka Iwasaki, Takaaki Ohkawauchi, Jun Ohya

    2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC)     1548 - 1551  2020.07  [Refereed]

    DOI

  • Locating Mechanical Switches Using RGB-D Sensor Mounted on a Disaster Response Robot

    Takuya Kanda, Kazuya Miyakawa, Jeonghwang Hayashi, Jun Ohya, Hiroyuki Ogata, Kenji Hashimoto, Xiao Sun, Takashi Matsuzawa, Hiroshi Naito, Atsuo Takanishi

    Electronic Imaging 2020, Intelligent Robotics and Industrial Applications using Computer Vision 2020   IRIACV ( 16 ) 1 - 6  2020.01  [Refereed]

  • Makeup Interpolation Based on Color and Shape Parametrization

    Jieun Cho, Jun Ohya, Sang Il Park

    Symmetry   11 ( 1422 ) 1 - 15  2019.11  [Refereed]

  • Disaster Response Robot’s Autonomous Manipulation of Valves in Disaster Sites Based on Visual Analyses of RGBD Images

    Keishi Nishikawa, Asaki Imai, Kazuya Miyakawa, Takuya Kanda, Takashi Matsuzawa, Kenji Hashimoto, Atsuo Takanishi, Hiroyuki Ogata, Jun Ohya

    2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)     4790 - 4797  2019.11  [Refereed]

  • Investigating Relationship between Running Motions and Skills Acquired from Jump Trainings,

    Chanjin Seo, Masato Sabanai, Hiroyuki Ogata, Jun Ohya

    icSPORTS     198 - 203  2019.09  [Refereed]

  • A Criterion that Evaluates the Sprinting Motion Using the Joint Coordinates Detected from the Video and Machine Learning

    Masato Sabanai, Chanjin Seo, Hiroyuki Ogata, Jun Ohya

    The 6th IIEEJ International Conference on Image Electronics and Visual Computing (IEVC 2019)   ( 3C-3 ) 1 - 4  2019.08  [Refereed]

  • Local Branch Ensemble Network: Autonomous Driving System Using End-to-End Deep Learning

    Zelin Zhang, Jun Ohya

    The 6th IIEEJ International Conference on Image Electronics and Visual Computing (IEVC 2019)   ( 2C-3 ) 1 - 4  2019.08  [Refereed]

  • Applying Curvatures Estimated from 3D Point Clouds to Environment Recognition in Forests Using SegNet

    Takeo Kaneko, Hiroyuki Fujimoto, Junya Morimoto, Takuya Hayashi, Junji Yamato, Jun Ohya

    The 6th IIEEJ International Conference on Image Electronics and Visual Computing (IEVC 2019)   ( 2C-4 ) 1 - 4  2019.08  [Refereed]

  • A Method of Proposing the Tracking Suitable for Supplementing an Exerciser’s Lacking Skill in Sprint

    Chanjin Seo, Masato Sabanai, Jun Ohya, Hiroyuki Ogata

    The 6th IIEEJ International Conference on Image Electronics and Visual Computing (IEVC 2019)   ( 3C-2 ) 1 - 4  2019.08  [Refereed]

  • Region Mining of Fetal Head in Ultrasound Image Based on Weakly Supervised Annotations and Deep Learning

    Yan LI, Rong XU, Artus KROHN-GRIMBERGHE, Jun OHYA, Hiroyasu IWATA

    IIEEJ Transactions on Image Electronics and Visual Computing   7 ( 1 ) 46 - 51  2019.06  [Refereed]

  • Deep Learning Based Uterus Localization and Anatomical Structure Segmentation on Fetal Ultrasound Image

    Yan LI, Rong XU, Artus KROHN-GRIMBERGHE, Jun OHYA, Hiroyasu IWATA

    IIEEJ Transactions on Image Electronics and Visual Computing   7 ( 1 ) 13 - 23  2019.06  [Refereed]

  • An Algorithm for Tracking Guitarists’ Fingertips Based on CNN-Segmentation and ROI Associated Particle Filter

    Zhao WANG, Jun OHYA

    Journal of Imaging Science and Technology   63 ( 2 ) 20506-1 - 20506-9  2019.03  [Refereed]

  • Detecting a Fetus in Ultrasound Images using Grad CAM and Locating the Fetus in the Uterus

    Genta Ishikawa, Rong Xu, Jun Ohya, Hiroyasu Iwata

    International Conference on Pattern Recognition Application and Methods     181 - 189  2019.02  [Refereed]

  • Understanding Sprinting Motion Skills Using Unsupervised Learning for Stepwise Skill Improvements of Running Motion

    Chanjin Seo, Masato Sabanai, Hiroyuki Ogata, Jun Ohya

    International Conference on Pattern Recognition Application and Methods     467 - 475  2019.02  [Refereed]

  • Detecting and Tracking Surgical Tools for Recognizing Phases of the Awake Brain Tumor Removal Surgery.

    Hiroki Fujie, Keiju Hirata, Takahiro Horigome, Hiroshi Nagahashi, Jun Ohya, Manabu Tamura, Ken Masamune, Yoshihiro Muragaki

    International Conference on Pattern Recognition Application and Methods     190 - 199  2019  [Refereed]

    DOI

  • Automatic Estimation of the Position and Orientation of the Drill to Be Grasped and Manipulated by the Disaster Response Robot Based on Analyzing Depth Camera Information

    Keishi Nishikawa, Jun Ohya, Hiroyuki Ogata, Kenji Hashimoto, Takashi Matsuzawa, Asaki Imai, Shunsuke Kimura, Atsuo Takanishi

    IS&T International Symposium on Electronic Imaging 2019, Intelligent Robotics and Industrial Applications using Computer Vision 2019   IRIACV ( 452 ) 1 - 6  2019.01  [Refereed]

  • Pattern and Frontier-based, Efficient and Effective Exploration of Autonomous Mobile Robots in Unknown Environments

    Hiroyuki FUJIMOTO, Junya MORIMOTO, Takuya HAYASHI, Junji YAMATO, Hiroyuki ISHII, Jun OHYA, Atsuo TAKANISHI

    IS&T International Symposium on Electronic Imaging 2019, Autonomous Vehicles and Machines Conference 2019     039-1 - 039-6  2019.01  [Refereed]

  • An Accurate and Robust Method for Detecting Fetal Heads in Ultrasound Images Based on Integrating a Voting Scheme and an Improved IRHT

    Genta Ishikawa, Rong Xu, Jun Ohya, Hiroyasu Iwata

    IIEEJ Transactions on Electronics and Visual Computing,   6 ( 2 ) 65 - 73  2018.12  [Refereed]

  • Automatic detection of valves with disaster response robot on basis of depth camera information

    Keishi Nishikawa, Jun Ohya, Hiroyuki Ogata, Kenji Hashimoto, Takashi Matsuzawa, Atsuo Takanishi

    Digital Image Computing: Techniques and Applications (DICTA 2018),     253 - 258  2018.12  [Refereed]

  • A 3D guitar fingering assessing system based on CNN-Hand pose estimation and SVR-Assessment

    Zhao Wang, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology   Part F138660   2781 - 2785  2018.01  [Refereed]

     View Summary

    © 2018, Society for Imaging Science and Technology. This paper proposes a guitar fingering assessing system based on CNN (Convolutional Neural Network) hand pose estimation and SVR (Support Vector Regression) evaluation. To spur our progress, first, a CNN architecture is proposed to estimate temporal 3D position of 16 joints of hand; then, based on a DCT (Discrete Cosine Transform) feature and SVR, fingering of guitarist is scored to interpret how well guitarist played. We also release a new dataset for professional guitar playing analysis with significant advantage in total number of video, professional judgement by expert of guitarist, accurate annotation for hand pose and score of guitar performance. Experiments using videos containing multiple persons' guitar plays under different conditions demonstrate that the proposed method outperforms the current state-of-art with (1) low mean error (Euclid distance of 6,1 mm) and high computation efficiency for hand pose estimation; (2) high rank correlation (0.68) for assessing the fingering (C major scale and symmetrical excise) of guitarists.

    DOI

  • Estimating the UAVs' crash point based on optical flows' voting in the image plane

    Yusaku Hatano, Hiroshi Nagahashi, Chen Yi, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology   Part F138654  2018.01  [Refereed]

     View Summary

    © 2018, Society for Imaging Science and Technology. Towards the actualization of an air bag system for the UAV's crash, this paper proposes a method for estimating the UAV's crash site from the video sequence acquired by the camera attached to the UAV. The crash point can be considered to correspond to the divergence point of the optical flows. In the accumulator, the cells at which the optical flows (straight lines) pass through are incremented by one. After performing this process for all the optical flows, the cell with the largest vote is obtained as the crash point (divergence point) in the image plane. Experiments using a hand held camera show that the accuracy of estimating the crash site is increased as the camera approaches the target plane. Overall, the experimental results are promising.

    DOI

  • An accurate and robust algorithm for tracking guitar neck in 3D based on modified RANSAC homography

    Zhao Wang, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology   Part F138651  2018.01  [Refereed]

     View Summary

    © 2018 Society for Imaging Science and Technology. Towards the actualization of an automatic guitar teaching system that can supervise guitar players, this paper proposes an algorithm for accurately and robustly tracking the 3D position of the fretboard from the video of guitar plays. First, we detect the SIFT features within the guitar fretboard and then match the detected points using KD-tree searching based matching algorithm frame by frame to track the whole fretboard. However, during the guitar plays, due to movements of the guitar neck or occlusions caused by guitar players' fingers, the feature points on the fretboard cannot always be matched accurately even though applying traditional RANSAC homography. Therefore, by using our modified RANSAC algorithm to filter out the matching error of the feature points, perspective transformation matrix is obtained between the correctly matched feature points detected at the first and other frames. Consequently, the guitar neck is tracked correctly based on the perspective transformation matrix. Experiments show promising results such as high accuracy: the total mean tracking error of only 4.17 mm and variance of 1.5 for the four tracked corners of the fretboard. This indicates the proposed method outperforms related tracking works including state-of-art Fully-convolutional Network.

    DOI

  • Discriminating the presence of the cerebral aneurysm using shape features obtained from medical images of the cerebral vessel

    Kosei Kikuchi, Takanobu Yagi, Xu Rong, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology   Part F138660   2831 - 2836  2018.01  [Refereed]

     View Summary

    © 2018, Society for Imaging Science and Technology. Towards the establishment of the preventive medical care for the cerebral aneurysm, this paper proposes an SVM based method for building a discrimination function that classifies the presence or absence of the cerebral aneurysm using the cerebral blood vessel's shape features obtained from medical images such as MR images. Using the discrimination function, this paper explores how much each feature affects the onset of the cerebral aneurysm. This paper deals with the internal carotid artery (ICA). The blood vessel (ICA)'s shape features are extracted from medical images of 18 persons without cerebral aneurysm and 13 patients with a cerebral aneurysm. From the medical image, the cross sections and centerline of the ICA are obtained. The cross sections are divided into nine sections along the centerline. Shape features such as the cross sectional area, its circularity, curvature, torsion, length of the centerline and branch angles are obtained in each section; as a total, 113 features including the mean and variance of some features in each section are used for building the SVM. As a result of conducting the experiments, the accuracy for discriminating the presence/absence of the aneurysm by the SVM is 90.3%. In the obtained discrimination function, the coefficient values of the function can be considered how much the features affect the onset of the aneurysm. The features that could significantly cause the onset of the cerebral aneurysm are clarified, and the reasons why these features are significant are discussed.

    DOI

  • An image processing based method for chewing detection using Variable-intensity template

    Atsuto Fujimoto, Takaaki Ohkawauchi, Junji Yamato, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology   Part F138660   2791 - 2794  2018.01  [Refereed]

     View Summary

    © 2018, Society for Imaging Science and Technology. To contribute to the actualization of the care worker assistance robot, this paper proposes a method for detecting whether the care receiver is chewing from the video sequence acquired by the camera that observes that receiver. The proposed method detects the receiver's face and areas for both cheeks and chin. After applying some normalization to the areas, chewing detection that uses a variable-intensity template is performed, where the template consists of shape models, interest points and intensity distribution model. A likelihood based on the variableintensity template is computed so that the receiver is judged whether the receiver is chewing. Experiments using seven subjects are conducted. As a result, the accuracy of chewing detection by the proposed method is 83%, which is quite promising.

    DOI

  • Exploring the effectiveness of using temporal order information for the early-recognition of suture surgery's six steps based on video image analyses of surgeons' hand actions

    Miwa Tsubota, Ye Li, Jun Ohya

    RO-MAN 2017 - 26th IEEE International Symposium on Robot and Human Interactive Communication   2017-   464 - 469  2017.12  [Refereed]

     View Summary

    To alleviate the recent shortage problem of nurses, the actualization of RSN (Robotic Scrub Nurse) that can autonomously judge the current step of the surgery and pass the surgical instruments needed for the next step to surgeons is desired. The authors developed a computer vision based algorithm that can early-recognize only two steps of suture surgery. Based on the past work, this paper explores the effectiveness of utilizing temporal order of the six steps in suture surgery for the early-recognition. Our early-recognition algorithm consists of two modules: start point detection and hand action early-recognition. Segments of the test video that start from each quasi-start point are compared with the training data, and their probabilities are calculated. According to the calculated probabilities, hand actions could be early-recognized. To improve the early-recognition accuracy, temporal order information could be useful. This paper checks confusions of three steps' early recognition results, and if necessary, early-recognizes again after eliminating the wrong result, while for the other three steps, temporal order information is not utilized. Experimental results show our early-recognition method that utilizes the temporal order information achieves better performances.

    DOI

  • Automatic fetal body and amniotic fluid segmentation from fetal ultrasound images by encoder-decoder network with inner layers

    Yan Li, Rong Xu, Jun Ohya, Hiroyasu Iwata

    Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS     1485 - 1488  2017.09  [Refereed]

     View Summary

    This paper explores the effectiveness of applying a deep learning based method to segment the amniotic fluid and fetal tissues in fetal ultrasound (US) images. The deeply learned model firstly encodes the input image into down scaled feature maps by convolution and pooling structures, then up-scale the feature maps to confidence maps by corresponded un-pooling and convolution layers. Additional convolution layers with 1×1 sized kernels are adopted to enhance the feature representations, which could be used to further improve the discriminative learning of our model. We effectively update the weights of the network by fine-tuning on part of the layers from a pre-trained model. By conducting experiments using clinical data, the feasibility of our proposed approach is compared and discussed. The result proves that this work achieves satisfied results for segmentation of specific anatomical structures from US images.

    DOI PubMed

  • Automatic, Accurate Estimation of the Position and Pose of a Ladder in 3D Point Cloud

    Keishi Nishikawa, Zhao Wang, Jun Ohya, Takashi Matsuzawa, Kenji Hashimoto, Atsuo Takanishi

    The IIEEJ Image Electronics and Visual Computing Workshop 2017   ( 5C-2 ) 1 - 6  2017.03  [Refereed]

  • An Accurate and Robust Fetal Head Detection Algorithm Integrating a Voting Scheme and an Improved IRHT Method

    Genta Ishikawa, Rong Xu, Jun Ohya, Hiroyasu Iwata

    The IIEEJ Image Electronics and Visual Computing Workshop 2017   ( 5C-3 ) 1 - 5  2017.03  [Refereed]

  • Pregnant Uterine Ultrasound Image Segmentation by Encoding-Decoding Convolutional Neural Network

    Yan Li, Rong Xu, Jun Ohya, Hiroyasu Iwata

    The IIEEJ Image Electronics and Visual Computing Workshop 2017   ( 1C-1 ) 1 - 4  2017.03  [Refereed]

  • Fingertips tracking algorithm for guitarist based on temporal grouping and pattern analysis

    Zhao Wang, Jun Ohya

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   10118   212 - 226  2017  [Refereed]

     View Summary

    This paper proposes a temporal Grouping and pattern analysis-based algorithm that could track the fingertips of guitarists during their guitar playing towards the actualization of the automatic guitar fingering recognition system. First a machine learning-based Bayesian Pixel Classifier is used to segment the hand area on the test data. Then, the probability map of fingertip is generated on the segmentation results by counting the voting numbers of the Template Matching and Reversed Hough Transform. Furthermore, a temporal Grouping algorithm, which is a geometry analysis for consecutive frames, is applied to removal noise and group the same fingertips (index finger, middle finger, ring finger, little finger). Then, a data association algorithm is utilized to associate 4 tracked fingers (index finger, middle finger, ring finger, little finger) with their correspondent tracked results frame by frame. Finally, particles are distributed only between the associated fingertip candidates to track the fingertips of guitarist effectively. The experimental result demonstrates that this fingertip tracking algorithm is robust enough for tracking fingertips (1) without any constrains such us color marker
    (2) under the complex contexts, such us complicated background, different illumination conditions, (3) with the high tracking accuracy (mean error 3.36 pixels for four fingertips).

    DOI

  • Fingertips tracking algorithm for guitarist based on temporal grouping and pattern analysis

    Zhao Wang, Jun Ohya

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   10118   212 - 226  2017  [Refereed]

     View Summary

    This paper proposes a temporal Grouping and pattern analysis-based algorithm that could track the fingertips of guitarists during their guitar playing towards the actualization of the automatic guitar fingering recognition system. First a machine learning-based Bayesian Pixel Classifier is used to segment the hand area on the test data. Then, the probability map of fingertip is generated on the segmentation results by counting the voting numbers of the Template Matching and Reversed Hough Transform. Furthermore, a temporal Grouping algorithm, which is a geometry analysis for consecutive frames, is applied to removal noise and group the same fingertips (index finger, middle finger, ring finger, little finger). Then, a data association algorithm is utilized to associate 4 tracked fingers (index finger, middle finger, ring finger, little finger) with their correspondent tracked results frame by frame. Finally, particles are distributed only between the associated fingertip candidates to track the fingertips of guitarist effectively. The experimental result demonstrates that this fingertip tracking algorithm is robust enough for tracking fingertips (1) without any constrains such us color marker
    (2) under the complex contexts, such us complicated background, different illumination conditions, (3) with the high tracking accuracy (mean error 3.36 pixels for four fingertips).

    DOI

  • Subaction Based Early Recognition of Surgeons’ Hand Actions from Continuous Surgery Videos

    Ye Li, Jun Ohya, Toshio Chiba, Xu Rong, Hiromasa Yamashita

    IIEEJ Transactions on Image Electronics and Visual Computing   4 ( 2 ) 124 - 135  2016.12  [Refereed]

  • Supervoxel-Based Detection and Segmentation of Moving Objects from Dynamic RGBD Images

    Naotomo Tatematsu, Jun Ohya, Larry S. Davis

    IIEEJ Transactions on Image Electronics and Visual Computing   4 ( 1 ) 32 - 42  2016.06  [Refereed]

  • Tracking the guitarist’s fingers as well as recognizing pressed chords from a video sequence

    Zhao Wang, Jun Ohya

    IS and T International Symposium on Electronic Imaging Science and Technology    2016  [Refereed]

     View Summary

    Towards the actualization of an autonomous guitar teaching system, this paper proposes the following two video analysis based methods: (1) pressed chord recognition and (2) fingertip tracking. For (1), an algorithm that can extract finger contours and chord changes is proposed so that the chords pressed by the guitar player are recognized. For (2), an algorithm that can track the fingertips by continuously monitoring the appearance and disappearance of the regions of fingertip candidates is proposed. Experimental results demonstrate that the proposed two modules are robust enough under complex contexts such as complicated background and different illumination conditions. Promising results were obtained for accurate tracking of fingertips and for accurate recognition of pressed chords.

    DOI

  • Detecting and Tracking the Guitar Neck Towards the Actualization of a Guitar Teaching-aid System

    Zhao Wang, Jun Ohya

    6th International Conference on Advanced Mechatronics (ICAM2015)     187 - 188  2015.12  [Refereed]

  • Estimating Clean-up Robots’ Mechanical Operations of Objects Using a SLAM Based Method

    Si Gao, Naotomo Tatematsu, Jun Ohya, Zhao Wang

    6th International Conference on Advanced Mechatronics (ICAM2015)     249 - 250  2015.12  [Refereed]

  • Early Recognition of Surgeons' Hand Actions from Continuous Surgery Videos

    Ye Li, Jun Ohya

    Interactive Medical Image Computing (IMIC), affiliated with MICCAI2015     78 - 85  2015.10  [Refereed]

  • Recognizing Hand Actions during Suture Surgeries from Video Sequences Based on Sliding Windows, SBoWS Feature and Neural Network

    Ye LI, Jun OHYA, Toshio CHIBA, Rong XU, Hiromasa YAMASHITA

    IIEEJ Transactions on Image Electronics and Visual Computing   3 ( 1 ) 34 - 43  2015.06  [Refereed]

  • Enhancing abstract imaginations of viewers of abstract paintings by a gaze based music generation system

    Tatsuya Ogusu, Jun Ohya, Jun Kurumisawa, Shunichi Yonemura

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   9173 ( LNCS 9173 ) 457 - 464  2015  [Refereed]

     View Summary

    The purpose of abstract painters is to let viewers get the various images and abstract images. However, viewers who do not have enough knowledge of art, cannot easily get abstract images. The authors have proposed a music generation system that utilizes viewers’ gazes. It can be expected that the authors’ music generation system can prompt the viewer of abstract paintings to imagine abstract images, which the painter intended to express. This paper explores whether the authors’ music generation system can enhance abstract imaginations of persons who see abstract paintings, by subjective tests. Experiments using 19 subjects and eight abstract paintings were conducted for the two cases in which the subjects see the abstract paintings without hearing any music and while hearing the viewers’ gaze based music generated by the authors’ system. Experimental results imply that “hearing gaze based music” could enhance the viewers’ abstract imagination.

    DOI

  • Enhancing abstract imaginations of viewers of abstract paintings by a gaze based music generation system

    Tatsuya Ogusu, Jun Ohya, Jun Kurumisawa, Shunichi Yonemura

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   9173   457 - 464  2015  [Refereed]

     View Summary

    The purpose of abstract painters is to let viewers get the various images and abstract images. However, viewers who do not have enough knowledge of art, cannot easily get abstract images. The authors have proposed a music generation system that utilizes viewers’ gazes. It can be expected that the authors’ music generation system can prompt the viewer of abstract paintings to imagine abstract images, which the painter intended to express. This paper explores whether the authors’ music generation system can enhance abstract imaginations of persons who see abstract paintings, by subjective tests. Experiments using 19 subjects and eight abstract paintings were conducted for the two cases in which the subjects see the abstract paintings without hearing any music and while hearing the viewers’ gaze based music generated by the authors’ system. Experimental results imply that “hearing gaze based music” could enhance the viewers’ abstract imagination.

    DOI

  • Detection and segmentation of moving objects from dynamic RGB and depth images

    Naotomo Tatematsu, Jun Ohya, Larry Davis

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   8971   19 - 34  2015  [Refereed]

     View Summary

    This paper proposes a method that combines Temporal Modified-RANSAC(TMR) with a fixation-based segmentation algorithm for reconstructing the 3D structure of moving and still objects that are tracked in video and depth image sequences acquired by moving Kinect© and/or range finders First, we compute 3D optical flow of feature points. Second, TMR classifies all flows into consistent 3D flow sets for the background and each moving object
    simultaneously, the rotation matrix and the translation vector for each 3D flow set are computed. Then, an improved fixation-based segmentation method segments each object’s area. Finally, dense 3D models for the background and each moving object are constructed along with each object’s rotation matrix and translation vector in each frame. Experiments using multiple moving objects in color and depth image sequences acquired by Kinect(c) demonstrate the effectiveness of our proposed method.

    DOI

  • Automatic Neural Network Based Generation of Line Drawing Based Portraits - Study of Conditions for Accurate Portrait Generation and Representing Noses -

    Makoto ENOMOTO, Guoyong FANG, Yuzhen WU, Jun OHYA

    Fourth IIEEJ International Workshop on Image Electronics and Visual Computing   ( 2C-3 ) 1 - 4  2014.10  [Refereed]

  • Study of Recognizing Customers’ Behaviors Based on Facial Expression Analysis by a Bayesian Network

    Luwei JIA, Jun OHYA, Takayuki HORI

    Fourth IIEEJ International Workshop on Image Electronics and Visual Computing   ( 2P-8 ) 1 - 4  2014.10  [Refereed]

  • A Bayesian Network Based Classification of Customers’ Behaviors in Stores Using Image Processing Based Extraction of Behavior Features

    Jing Yan, Takayuki Hori, Jun Ohya

    Fourth IIEEJ International Workshop on Image Electronics and Visual Computing   ( 4A-4 ) 1 - 4  2014.10  [Refereed]

  • Locating Fetal Facial Surface, Oral Cavity and Airways by a 3D Ultrasound Calibration Using a Novel Cones' Phantom

    Rong Xu, Jun Ohya, Yoshinobu Sato, Bo Zhang, Masakatsu G. Fujie

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E97D ( 5 ) 1324 - 1335  2014.05  [Refereed]

     View Summary

    Toward the actualization of an automatic navigation system for fetoscopic tracheal occlusion (FETO) surgery, this paper proposes a 3D ultrasound (US) calibration-based approach that can locate the fetal facial surface, oral cavity, and airways by a registration between a 3D fetal model and 3D US images. The proposed approach consists of an offline process and online process. The offline process first reconstructs the 3D fetal model with the anatomies of the oral cavity and airways. Then, a point-based 3D US calibration system based on real-time 3D US images, an electromagnetic (EM) tracking device, and a novel cones' phantom, computes the matrix that transforms the 3D US image space into the world coordinate system. In the online process, by scanning the mother's body with a 3D US probe, 3D US images containing the fetus are obtained. The fetal facial surface extracted from the 3D US images is registered to the 3D fetal model using an ICP-based (iterative closest point) algorithm and the calibration matrices, so that the fetal facial surface as well as the oral cavity and airways are located. The results indicate that the 3D US calibration system achieves an FRE (fiducial registration error) of 1.49 +/- 0.44 mm and a TRE (target registration error) of 1.81 +/- 0.56 mm by using 24 fiducial points from two US volumes. A mean TRE of 1.55 +/- 0.46 mm is also achieved for measuring location accuracy of the 3D fetal facial surface extracted from 3D US images by 14 target markers, and mean location errors of 2.51 +/- 0.47 mm and 3.04 +/- 0.59 mm are achieved for indirectly measuring location accuracy of the pharynx and the entrance of the trachea, respectively, which satisfy the requirement of the FETO surgery.

    DOI

  • 相異なる照明光下の固定カメラ画像からの照明光色推定法

    川村春美, 八尾泰洋, 米村俊一, 大谷 淳, 小島 明

    画像電子学会誌   43 ( 2 ) 164 - 174  2014.03  [Refereed]

    DOI J-GLOBAL

  • Recognizing Surgeon's Actions during Suture Operations from Video Sequences

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    MEDICAL IMAGING 2014: IMAGE PROCESSING   9034  2014  [Refereed]

     View Summary

    Because of the shortage of nurses in the world, the realization of a robotic nurse that can support surgeries autonomously is very important. More specifically, the robotic nurse should be able to autonomously recognize different situations of surgeries so that the robotic nurse can pass necessary surgical tools to the medical doctors in a timely manner. This paper proposes and explores methods that can classify suture and tying actions during suture operations from the video sequence that observes the surgery scene that includes the surgeon's hands.
    First, the proposed method uses skin pixel detection and foreground extraction to detect the hand area. Then, interest points are randomly chosen from the hand area so that their 3D SIFT descriptors are computed. A word vocabulary is built by applying hierarchical K-means to these descriptors, and the words' frequency histogram, which corresponds to the feature space, is computed Finally, to classify the actions, either SVM (Support Vector Machine), Nearest Neighbor rule (NN) for the feature space or a method that combines "sliding window" with NN is performed.
    We collect 53 suture videos and 53 tying videos to build the training set and to test the proposed method experimentally It turns out that the NN gives higher than 90% accuracies, which are better recognition than SVM. Negative actions, which are different from either suture or tying action, are recognized with quite good accuracies, while "Sliding window" did not show significant improvements for suture and tying and cannot recognize negative actions.

    DOI

  • Automatic Recognition of Square Notation Symbols in Western Plainchant Manuscripts

    Carolina Ramirez, Jun Ohya

    JOURNAL OF NEW MUSIC RESEARCH   43 ( 4 ) 390 - 399  2014  [Refereed]

     View Summary

    While the Optical Music Recognition (OMR) of printed and handwritten music scores in modern standard notation has been broadly studied, this is not the case for early music manuscripts. This is mainly due to the high variability in the sources introduced by their severe physical degradation, the lack of notation standards and, in the case of the scanned versions, by non-homogenous image-acquisition protocols. The volume of early musical manuscripts available is considerable, and therefore we believe that computational methods can be extremely useful in helping to preserve, share and analyse this information. This paper presents an approach to recognizing handwritten square musical notation in degraded western plainchant manuscripts from the XIVth to XVIth centuries. We propose the use of image processing techniques that behave robustly under high data variability and which do not require strong hypotheses regarding the condition of the sources. The main differences from traditional OMR approaches are our avoidance of the staff line removal stage and the use of grey-level images to perform primitive segmentation and feature extraction. We used 136 images from the Digital Scriptorium repository (DS, 2007), from which we were able to extract over 90% of the staves and over 88% of all symbols present. For symbol classification, we used gradient-based features and SVM classifiers, obtaining over 90% precision and recall over eight basic symbol classes.

    DOI

  • Recognizing surgeon's actions during suture operations from video sequences

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    Progress in Biomedical Optics and Imaging - Proceedings of SPIE   9034  2014  [Refereed]

     View Summary

    Because of the shortage of nurses in the world, the realization of a robotic nurse that can support surgeries autonomously is very important. More specifically, the robotic nurse should be able to autonomously recognize different situations of surgeries so that the robotic nurse can pass necessary surgical tools to the medical doctors in a timely manner. This paper proposes and explores methods that can classify suture and tying actions during suture operations from the video sequence that observes the surgery scene that includes the surgeon's hands. First, the proposed method uses skin pixel detection and foreground extraction to detect the hand area. Then, interest points are randomly chosen from the hand area so that their 3D SIFT descriptors are computed. A word vocabulary is built by applying hierarchical K-means to these descriptors, and the words frequency histogram, which corresponds to the feature space, is computed. Finally, to classify the actions, either SVM (Support Vector Machine), Nearest Neighbor rule (NN) for the feature space or a method that combines sliding window with NN is performed. We collect 53 suture videos and 53 tying videos to build the training set and to test the proposed method experimentally. It turns out that the NN gives higher than 90% accuracies, which are better recognition than SVM. Negative actions, which are different from either suture or tying action, are recognized with quite good accuracies, while Sliding window did not show significant improvements for suture and tying and cannot recognize negative actions. © 2014 SPIE.

    DOI

  • Inspiring Viewers of Abstract Painting by a Gaze Based Music Generation

    Tatsuya Ogusu, Jun Ohya, Jun Kurumisawa, Shunichi Yonemura

    Communications in Computer and Information Science   434   213 - 217  2014  [Refereed]

     View Summary

    This paper explores the effectiveness of prompting abstract paintings' viewers' inspiration and imagination by the authors' gaze based music generation system. The authors' music generation system detects the viewer's gaze by a gaze detection equipment. At each of the gaze staying positions in the painting, the color of that point is converted to the sound so that as the gaze moves, music that consists of the converted time series sounds is generated. Experiments using six subjects and six abstract paintings were conducted for the three cases in which the subjects see the abstract paintings without hearing any music, while hearing pre-selected music and while hearing the viewers' gaze based music generated by the authors' system. The experimental results imply that "hearing gaze based music" could stimulate the viewers' inspiration and imagination best, "hearing pre-selected music" second best, and "without music" third best. © Springer International Publishing Switzerland 2014.

    DOI

  • Locating fetal facial surface, oral cavity and airways by a 3D ultrasound calibration using a novel cones' phantom

    Rong Xu, Jun Ohya, Yoshinobu Sato, Bo Zhang, Masakatsu G. Fujie

    IEICE Transactions on Information and Systems   E96-D ( 5 ) 1324 - 1335  2014  [Refereed]

     View Summary

    Toward the actualization of an automatic navigation system for fetoscopic tracheal occlusion (FETO) surgery, this paper proposes a 3D ultrasound (US) calibration-based approach that can locate the fetal facial surface, oral cavity, and airways by a registration between a 3D fetal model and 3D US images. The proposed approach consists of an offline process and online process. The offline process first reconstructs the 3D fetal model with the anatomies of the oral cavity and airways. Then, a point-based 3D US calibration system based on real-time 3D US images, an electromagnetic (EM) tracking device, and a novel cones' phantom, computes the matrix that transforms the 3D US image space into the world coordinate system. In the online process, by scanning the mother's body with a 3D US probe, 3D US images containing the fetus are obtained. The fetal facial surface extracted from the 3D US images is registered to the 3D fetal model using an ICP-based (iterative closest point) algorithm and the calibration matrices, so that the fetal facial surface as well as the oral cavity and airways are located. The results indicate that the 3D US calibration system achieves an FRE (fiducial registration error) of 1.49±0.44mm and a TRE (target registration error) of 1.81±0.56mm by using 24 fiducial points from two US volumes. A mean TRE of 1.55±0.46 mm is also achieved for measuring location accuracy of the 3D fetal facial surface extracted from 3D US images by 14 target markers, and mean location errors of 2.51±0.47 mm and 3.04±0.59 mm are achieved for indirectly measuring location accuracy of the pharynx and the entrance of the trachea, respectively, which satisfy the requirement of the FETO surgery. © 2014 The Institute of Electronics, Information and Communication Engineers.

    DOI

  • Automatic Lip Reading by Using Multimodal Visual Features

    Shohei Takahashi, Jun Ohya

    INTELLIGENT ROBOTS AND COMPUTER VISION XXXI: ALGORITHMS AND TECHNIQUES   9025  2014  [Refereed]

     View Summary

    Since long time ago, speech recognition has been researched, though it does not work well in noisy places such as in the car or in the train. In addition, people with hearing-impaired or difficulties in hearing cannot receive benefits from speech recognition. To recognize the speech automatically, visual information is also important. People understand speeches from not only audio information, but also visual information such as temporal changes in the lip shape. A vision based speech recognition method could work well in noisy places, and could be useful also for people with hearing disabilities.
    In this paper, we propose an automatic lip-reading method for recognizing the speech by using multimodal visual information without using any audio information such as speech recognition. First, the ASM (Active Shape Model) is used to track and detect the face and lip in a video sequence. Second, the shape, optical flow and spatial frequencies of the lip features are extracted from the lip detected by ASM. Next, the extracted multimodal features are ordered chronologically so that Support Vector Machine is performed in order to learn and classify the spoken words. Experiments for classifying several words show promising results of this proposed method.

    DOI

  • Quasi Random Scheme Based Improved Iterative Randomized Hough Transform (IRHT) for Detection of Partial Ellipses in Medical Images

    Rong XU, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    IIEEJ Transactions on Image Electronics and Visual Computing   1 ( 1 ) 97 - 106  2013.12  [Refereed]

  • 高彩度・低彩度gamut と反対色特性を利用した灰色仮説ベースの照明光色推定法

    川村春美, 米村俊一, 大谷 淳, 小島 明

    電子情報通信学会論文誌   J96-D ( 12 ) 3079 - 3089  2013.12  [Refereed]

    CiNii

  • Towards Recognizing Surgeon's Action during Suture Surgery from the Video Sequence

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    The 9th Asian Conference on Computer Aided Surgery     66 - 67  2013.09  [Refereed]

  • Automatic Real-time Tracking of Fetal Mouth in Fetoscopic Video Sequence for Supporting Fetal Surgeries

    Rong Xu, Tianliang Xie, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    MEDICAL IMAGING 2013: IMAGE-GUIDED PROCEDURES, ROBOTIC INTERVENTIONS, AND MODELING   8671   1 - 7  2013  [Refereed]

     View Summary

    Recently, a minimally invasive surgery (MIS) called fetoscopic tracheal occlusion (FETO) was developed to treat severe congenital diaphragmatic hernia (CDH) via fetoscopy, by which a detachable balloon is placed into the fetal trachea for preventing pulmonary hypoplasia through increasing the pressure of the chest cavity. This surgery is so dangerous that a supporting system for navigating surgeries is deemed necessary. In this paper, to guide a surgical tool to be inserted into the fetal trachea, an automatic approach is proposed to detect and track the fetal face and mouth via fetoscopic video sequencing. More specifically, the AdaBoost algorithm is utilized as a classifier to detect the fetal face based on Haar-like features, which calculate the difference between the sums of the pixel intensities in each adjacent region at a specific location in a detection window. Then, the CamShift algorithm based on an iterative search in a color histogram is applied to track the fetal face, and the fetal mouth is fitted by an ellipse detected via an improved iterative randomized Hough transform approach. The experimental results demonstrate that the proposed automatic approach can accurately detect and track the fetal face and mouth in real-time in a fetoscopic video sequence, as well as provide an effective and timely feedback to the robot control system of the surgical tool for FETO surgeries.

    DOI

  • Gray-world-assumption-based illuminant color estimation using color gamuts with high and low chroma

    Harumi Kawamura, Shunichi Yonemura, Jun Ohya, Akira Kojima

    Proceedings of SPIE - The International Society for Optical Engineering   8652 ( 86520C ) 1 - 8  2013  [Refereed]

     View Summary

    A new approach is proposed for estimating illuminant colors from color images under an unknown scene illuminant. The approach is based on a combination of a gray-world-assumption-based illuminant color estimation method and a method using color gamuts. The former method, which is one we had previously proposed, improved on the original method that hypothesizes that the average of all the object colors in a scene is achromatic. Since the original method estimates scene illuminant colors by calculating the average of all the image pixel values, its estimations are incorrect when certain image colors are dominant. Our previous method improves on it by choosing several colors on the basis of an opponent-color property, which is that the average color of opponent colors is achromatic, instead of using all colors. However, it cannot estimate illuminant colors when there are only a few image colors or when the image colors are unevenly distributed in local areas in the color space. The approach we propose in this paper combines our previous method and one using high chroma and low chroma gamuts, which makes it possible to find colors that satisfy the gray world assumption. High chroma gamuts are used for adding appropriate colors to the original image and low chroma gamuts are used for narrowing down illuminant color possibilities. Experimental results obtained using actual images show that even if the image colors are localized in a certain area in the color space, the illuminant colors are accurately estimated, with smaller estimation error average than that generated in the conventional method. © 2013 SPIE.

    DOI

  • A flexible surgical tool localization using a 3D ultrasound calibration system for fetoscopic tracheal occlusion (FETO)

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   7761   17 - 24  2013  [Refereed]

     View Summary

    In fetoscopic tracheal occlusion (FETO) surgery, the localization of a flexible tool has an important role in easing high demands for surgeons. In this paper, a point-based 3D ultrasound (US) calibration system is proposed for localizing the surgical tool, based on real-time 3D US images, an electromagnetic (EM) tracking device, and a novel cones' phantom. A 3D US probe attached with an EM sensor is used to acquire the 3D US images of the surgical tool
    meanwhile, an EM transmitter is fixed at a certain point in the environment as the origin of the world coordinate system. From these 3D US images, the skeleton of the surgical tool is extracted, and 15 labeled points along the surgical tool are then distinguished, whose coordinates in the world coordinate system are calculated by the proposed 3D US calibration system. The results demonstrate that our proposed approach achieves high localization accuracy, i.e. the average TRE of 1.99 ± 0.49 mm and the maximum TRE of 2.61 ± 0.49 mm, which satisfy the requirement of the FETO surgery: less than 3.85 mm. © Springer-Verlag 2013.

    DOI

  • Automatic Fetal Face Detection By Locating Fetal Facial Features From 3D Ultrasound Images For Navigating Fetoscopic Tracheal Occlusion Surgeries

    Rong Xu, Jun Ohya, Bo Zhang, Masakatsu G. Fujie, Yoshinobu Sato

    2013 10TH IEEE INTERNATIONAL CONFERENCE AND WORKSHOPS ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG)   ( 104 ) 1 - 6  2013  [Refereed]

     View Summary

    With the wide clinical application of 3D ultrasound (US) imaging, automatic location of fetal facial features from US volumes for navigating fetoscopic tracheal occlusion (FETO) surgeries becomes possible, which plays an important role in reducing surgical risk. In this paper, we propose a feature-based method to automatically detect 3D fetal face and accurately locate key facial features without any priori knowledge or training data. The candidates of the key facial features, such as the nose, eyes, nose upper bridge and upper lip are detected by analyzing the mean and Gaussian curvatures of the facial surface. Each feature is gradually identified from the candidates by a boosting traversal scheme based on the spatial relations between each feature. In experiments, all key feature points are detected for each case, and thus a detection success rate of 100% is achieved by using 72 3D US images from a test database of 6 fetal faces in the frontal view and any pose within 15 degrees from the frontal view, and the location error 3.18 +/- 0.91 inn of the detected upper lip for all test data is obtained, which can be tolerated by the FETO surgery. Moreover, this system has a high efficiency and can detect all key facial features in about 625 ms on a quadcore 2.60 GHz computer.

  • An investigation into encouraging higher user self-disclosure during text-based computer mediated communication: A case study of prompter communication system

    Li Jen Chen, Mutsumi Suganuma, Shigekazu Sakai, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    Journal of the Institute of Image Electronics Engineers of Japan   42 ( 1 ) 89 - 101  2013.01  [Refereed]

     View Summary

    An investigation into encouraging higher user self-disclosure during text-based Computer Mediated Communication (CMC) system known as the Prompter Communication system (PCS), proposed by Yonemura et al. in previous studies, in which 1 to 3 black circular prompter images ("•") are embedded onto the background of a text-field canvas as stimuli and as an embryonic basis for conversation, have been investigated and evaluated. The effects on supporting the projection or reflection of generated mental representations onto messages with the "•" prompter image as part of the representations within the message contents have also been examined. During the investigation, the concepts and features of the Geneplore cognitive creative thinking process model, together with self-projections, were integrated into the proposed system. Based on these concepts and features, the PCS aims to encourage users to express their mental representations by projecting or reflecting them onto the messages in combination with the provided "•" prompter image as part of the message contents. Communication experiments using the prototype PCS system have been conducted, where a Normal Communication System, NCS which uses only a plain text-field canvas without the "•" prompter image, is compared with the PCS. The experimental results indicate that communication conducted using the PCS results in an increased number of messages with the "•" prompter image(s) involved as part of the message contents than as background images. Further analysis into the self-disclosure level of the message contents and the relationships between the degree of self-disclosure and the representations of the "•" prompter image(s) within the messages have indicated that the degrees of self-disclosure increases when the "•" prompter image(s) are used to represent the participant him/herself, or the participant's thoughts and emotions, and decreases when the "•" prompter image(s) are used to represent animals or objects, or when the "•" prompter image(s) presented within the messages have been ignored. Investigations into the differences in self-disclosure level between message contents from both the PCS and the NCS have shown that using the PCS results in a generally higher degree in the average frequency of the appearance of high self-disclosure information than when using the NCS. During the conversation, these messages contain information with a high level of self-disclosure, which, consequently, may further lead to the development of intimate and deep interpersonal relationships.

    DOI

  • A flexible surgical tool localization using a 3D ultrasound calibration system for fetoscopic tracheal occlusion (FETO)

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)   7761   17 - 24  2013  [Refereed]

     View Summary

    In fetoscopic tracheal occlusion (FETO) surgery, the localization of a flexible tool has an important role in easing high demands for surgeons. In this paper, a point-based 3D ultrasound (US) calibration system is proposed for localizing the surgical tool, based on real-time 3D US images, an electromagnetic (EM) tracking device, and a novel cones' phantom. A 3D US probe attached with an EM sensor is used to acquire the 3D US images of the surgical tool
    meanwhile, an EM transmitter is fixed at a certain point in the environment as the origin of the world coordinate system. From these 3D US images, the skeleton of the surgical tool is extracted, and 15 labeled points along the surgical tool are then distinguished, whose coordinates in the world coordinate system are calculated by the proposed 3D US calibration system. The results demonstrate that our proposed approach achieves high localization accuracy, i.e. the average TRE of 1.99 ± 0.49 mm and the maximum TRE of 2.61 ± 0.49 mm, which satisfy the requirement of the FETO surgery: less than 3.85 mm. © Springer-Verlag 2013.

    DOI

  • Gray-world-assumption-based illuminant color estimation using color gamuts with high and low chroma

    Harumi Kawamura, Shunichi Yonemura, Jun Ohya, Akira Kojima

    Proceedings of SPIE - The International Society for Optical Engineering   8652  2013  [Refereed]

     View Summary

    A new approach is proposed for estimating illuminant colors from color images under an unknown scene illuminant. The approach is based on a combination of a gray-world-assumption-based illuminant color estimation method and a method using color gamuts. The former method, which is one we had previously proposed, improved on the original method that hypothesizes that the average of all the object colors in a scene is achromatic. Since the original method estimates scene illuminant colors by calculating the average of all the image pixel values, its estimations are incorrect when certain image colors are dominant. Our previous method improves on it by choosing several colors on the basis of an opponent-color property, which is that the average color of opponent colors is achromatic, instead of using all colors. However, it cannot estimate illuminant colors when there are only a few image colors or when the image colors are unevenly distributed in local areas in the color space. The approach we propose in this paper combines our previous method and one using high chroma and low chroma gamuts, which makes it possible to find colors that satisfy the gray world assumption. High chroma gamuts are used for adding appropriate colors to the original image and low chroma gamuts are used for narrowing down illuminant color possibilities. Experimental results obtained using actual images show that even if the image colors are localized in a certain area in the color space, the illuminant colors are accurately estimated, with smaller estimation error average than that generated in the conventional method. © 2013 SPIE.

    DOI

  • Automatic Real-time Tracking of Fetal Mouth in Fetoscopic Video Sequence for Supporting Fetal Surgeries

    Rong Xu, Tianliang Xie, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    MEDICAL IMAGING 2013: IMAGE-GUIDED PROCEDURES, ROBOTIC INTERVENTIONS, AND MODELING   8671  2013  [Refereed]

     View Summary

    Recently, a minimally invasive surgery (MIS) called fetoscopic tracheal occlusion (FETO) was developed to treat severe congenital diaphragmatic hernia (CDH) via fetoscopy, by which a detachable balloon is placed into the fetal trachea for preventing pulmonary hypoplasia through increasing the pressure of the chest cavity. This surgery is so dangerous that a supporting system for navigating surgeries is deemed necessary. In this paper, to guide a surgical tool to be inserted into the fetal trachea, an automatic approach is proposed to detect and track the fetal face and mouth via fetoscopic video sequencing. More specifically, the AdaBoost algorithm is utilized as a classifier to detect the fetal face based on Haar-like features, which calculate the difference between the sums of the pixel intensities in each adjacent region at a specific location in a detection window. Then, the CamShift algorithm based on an iterative search in a color histogram is applied to track the fetal face, and the fetal mouth is fitted by an ellipse detected via an improved iterative randomized Hough transform approach. The experimental results demonstrate that the proposed automatic approach can accurately detect and track the fetal face and mouth in real-time in a fetoscopic video sequence, as well as provide an effective and timely feedback to the robot control system of the surgical tool for FETO surgeries.

    DOI

  • Automatic Fetal Face Detection By Locating Fetal Facial Features From 3D Ultrasound Images For Navigating Fetoscopic Tracheal Occlusion Surgeries

    Rong Xu, Jun Ohya, Bo Zhang, Masakatsu G. Fujie, Yoshinobu Sato

    2013 10TH IEEE INTERNATIONAL CONFERENCE AND WORKSHOPS ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG)    2013  [Refereed]

     View Summary

    With the wide clinical application of 3D ultrasound (US) imaging, automatic location of fetal facial features from US volumes for navigating fetoscopic tracheal occlusion (FETO) surgeries becomes possible, which plays an important role in reducing surgical risk. In this paper, we propose a feature-based method to automatically detect 3D fetal face and accurately locate key facial features without any priori knowledge or training data. The candidates of the key facial features, such as the nose, eyes, nose upper bridge and upper lip are detected by analyzing the mean and Gaussian curvatures of the facial surface. Each feature is gradually identified from the candidates by a boosting traversal scheme based on the spatial relations between each feature. In experiments, all key feature points are detected for each case, and thus a detection success rate of 100% is achieved by using 72 3D US images from a test database of 6 fetal faces in the frontal view and any pose within 15 degrees from the frontal view, and the location error 3.18 +/- 0.91 inn of the detected upper lip for all test data is obtained, which can be tolerated by the FETO surgery. Moreover, this system has a high efficiency and can detect all key facial features in about 625 ms on a quadcore 2.60 GHz computer.

    DOI

  • 3D Surface Reconstruction Based on Image Stitching from Gastric Endoscopic Video Sequence

    Mengyao Duan, Rong Xu, Jun Ohya

    APPLICATIONS OF DIGITAL IMAGE PROCESSING XXXVI   8856  2013  [Refereed]

     View Summary

    This paper proposes a method for reconstructing 3D detailed structures of internal organs such as gastric wall from endoscopic video sequences. The proposed method consists of the four major steps: Feature-point-based 3D reconstruction, 3D point cloud stitching, dense point cloud creation and Poisson surface reconstruction. Before the first step, we partition one video sequence into groups, where each group consists of two successive frames (image pairs), and each pair in each group contains one overlapping part, which is used as a stitching region. Fist, the 3D point cloud of each group is reconstructed by utilizing structure from motion (SFM). Secondly, a scheme based on SIFT features registers and stitches the obtained 3D point clouds, by estimating the transformation matrix of the overlapping part between different groups with high accuracy and efficiency. Thirdly, we select the most robust SIFT feature points as the seed points, and then obtain the dense point cloud from sparse point cloud via a depth testing method presented by Furukawa. Finally, by utilizing Poisson surface reconstruction, polygonal patches for the internal organs are obtained. Experimental results demonstrate that the proposed method achieves a high accuracy and efficiency for 3D reconstruction of gastric surface from an endoscopic video sequence.

    DOI

  • e-learning 用講義ビデオにおける 学習者の学習行動を利用した主観的難易度の把握方法の基礎的検討

    大川内隆朗, 大谷淳, 米村俊一, 徳永幸生

    日本教育工学会論文誌   36 ( 3 ) 193 - 203  2012.12  [Refereed]

    DOI J-GLOBAL

  • Study of a teacher support method for improving lectures for e-learning on demand

    Takaaki Ohkawauchi, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    IEVC (IIEEJ 3rd Image Electronics and Visual Computing Workshop) 2012   ( 2A-4 ) 1 - 4  2012.11  [Refereed]

  • Study of generating portraits using neural networks - Portraits that consist of facial contour, eyebrows, eyes and mouth ?

    Makoto Enomoto, Akira Takano, Jun Ohya

    IEVC (IIEEJ 3rd Image Electronics and Visual Computing Workshop) 2012   ( 4A-4 ) 1 - 4  2012.11  [Refereed]

  • Proposed Prompter Communication System for Supporting Non-verbal Expressions in order to Encourage Development of Deep and Intimate Interpersonal Relationships

    Chen Li Jen, Mutsumi Suganuma, Shigekazu Sakai, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    早稲田大学GITS/GITI紀要2011-2012     20 - 30  2012.10  [Refereed]

  • e-learning用講義コンテンツにおける教員の講義改善活動を支援するシステムの提案と評価

    大川内隆朗, 大谷淳, 米村俊一, 徳永幸生

    早稲田大学GITS/GITI紀要   2011-2012   11 - 19  2012.10  [Refereed]

    J-GLOBAL

  • Detection and Segmentation of Moving Objects from Dynamic RGB and Depth Images

    Naotomo Tatematsu, Jun Ohya, Larry Davis

    3D Camera Workshop in conjunction with International Conference on Virtual Reality and Visualization     1 - 8  2012.09  [Refereed]

  • Estimating object colors in outdoor scenes with same object region but different illuminants

    Harumi Kawamura, Yasuhiro Yao, Shunichi Yonemura, Jun Ohya, Akira Kojima

    AIC (The International Colour Association) 2012 Interim Meeting     582 - 585  2012.09  [Refereed]

  • Generating Various Face Makeup Styles based on a Small Set of Examples

    Jieun Cho, Minje Park, Sang Il Park, Jun Ohya

    CGI (Computer Graphics International) 2012   ( 222 ) 1 - 4  2012.06  [Refereed]

  • Segmentation of Brain MRI

    Rong Xu, Limin Luo, Jun Ohya

    Advances in Brain Imaging     143 - 170  2012.02

  • Tracking White Road Line by Particle Filter from the Video Sequence Acquired by the Camera Attached to a Walking Human Body

    Shohei Takahashi, Jun Ohya

    IMAGE PROCESSING: ALGORITHMS AND SYSTEMS X AND PARALLEL PROCESSING FOR IMAGING APPLICATIONS II   8295   82950 - 82959  2012  [Refereed]

     View Summary

    This paper proposes a method for tracking and recognizing the white line marked in the surface of the road from the video sequence acquired by the camera attached to a walking human, towards the actualization of an automatic navigation system for the visually handicapped. Our proposed method consists of two main modules: (1) Particle Filter based module for tracking the white line, and (2) CLAFIC Method based module for classifying whether the tracked object is the white line. In (1), each particle is a rectangle, and is described by its centroid's coordinates and its orientation. The likelihood of a particle is computed based on the number of white pixels in the rectangle. In (2), in order to obtain the ranges (to be used for the recognition) for the white line's length and width, Principal Component Analysis is applied to the covariance matrix obtained from valid sample particles. At each frame, PCA is applied to the covariance matrix constructed from particles with high likelihood, and if the obtained length and width are within the above-mentioned ranges, it is recognized as the white line. Experimental results using real video sequences show the validity of the proposed method.

    DOI

  • Accurate, Dense 3D Reconstruction of Moving and Still Objects from Dynamic Stereo Sequences Based on Temporal Modified-RANSAC and Feature-Cut

    Naotomo Tatematsu, Jun Ohya

    INTELLIGENT ROBOTS AND COMPUTER VISION XXIX: ALGORITHMS AND TECHNIQUES   8301   830105 - 830115  2012  [Refereed]

     View Summary

    This paper improves the authors' conventional method for reconstructing the 3D structure of moving and still objects that are tracked in the video and/or depth image sequences acquired by moving cameras and/or range finder. The authors proposed a Temporal Modified-RANSAC based method [1] that (1) can discriminate each moving object from the still background in color image and depth image sequences acquired by moving stereo cameras or moving range finder, (2) can compute the stereo cameras' egomotion, (3) can compute the motion of each moving object, and (4) can reconstruct the 3D structure of each moving object and the background. However, the TMR based method has the following two problems concerning the 3D reconstruction: lack of accuracy of segmenting into each object's region and sparse 3D reconstructed points in each object's region. To solve these problems of our conventional method, this paper proposes a new 3D segmentation method that utilizes Graph-cut, which is frequently used for segmentation tasks. First, the proposed method tracks feature points in the color and depth image sequences so that 3D optical flows of the feature points in every N frames are obtained. Then, TMR classifies all the obtained 3D optical flows into regions (3D flow set) for the background and each moving object; simultaneously, the rotation matrix and the translation vector for each 3D flow set are computed. Next, Graph-Cut using the energy function that consists of color probability, structure probability and a-priori probability is performed so that pixels in each frame are segmented into object regions and the background region. Finally, 3D point clouds are obtained from the segmentation result image and depth image, and then the point clouds are merged using the rotation and translation from the N-th frame prior to the current frame so that 3D models for the background and each moving object are constructed with dense 3D point data.

    DOI

  • Human Gesture Analysis using Multimodal features

    Luo Dan, Hazim Kemal Ekenel, Ohya Jun

    2012 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW)     471 - 476  2012  [Refereed]

     View Summary

    Human gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and PLS is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.

    DOI

  • Human Gesture Analysis using Multimodal features

    Luo Dan, Hazim Kemal Ekenel, Ohya Jun

    2012 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW)     471 - 476  2012  [Refereed]

     View Summary

    Human gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and PLS is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.

    DOI

  • Automatic Fetal Head Detection on Ultrasound Images by An Improved Iterative Randomized Hough Transform

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    Proceedings of 26th International Conference of Image and Vision Computing New Zealand (IVCNZ’ 2011)     288 - 292  2011.11  [Refereed]

  • Visualizing Large Procedural Volumetric Terrains Using Nested Clip-Boxes

    Sven Forstmann, Jun Ohya

    GITS/GITI 紀要     51 - 61  2011.10  [Refereed]

  • 反対色の特性を利用した色選択による灰色仮説判定に基づくカラー画像からの照明光推定法

    川村春美, 米村俊一, 大谷 淳, 松浦宣彦

    電子情報通信学会論文誌 D   J94-D ( 8 ) 1346 - 1358  2011.08  [Refereed]

    J-GLOBAL

  • Facial Analysis Aided Human Gesture Recognition for Human Computer Interaction

    Dan Luo, Hua Gao, Hazim Kemal Ekenel, Jun Ohya

    Proceedings of MVA2011 (IAPR Conference on Machine Vision Applications)     446 - 449  2011.06  [Refereed]

  • Conversation strategies and characteristics in line drawing video communication system

    Shun-Ichi Yonemurat, Ken Hanada, Yukio Tokunaga, Satoshi Shimada, Akira Suzuki, Jun Ohya

    Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers   65 ( 4 ) 515 - 522  2011.04  [Refereed]

     View Summary

    The characteristics of line drawing video communication are described in this paper. In the experiment, an original video and a line drawing video were compared. Subjects performed the same task using both video systems and their performances were assessed in terms of the number of utterances in the conversation and the subject's switching pause. The subject's conversation strategy was extracted based on gaze measurement. In the case of the subujects who often gaze at a partner's eyes during conversation, the switching pause for the line drawing video was significantly longer than for the original video. In both video formats, subjects who does not gaze at a partner's eyes had more switching pauses and utterances than the subjects who do.

    DOI

  • Illuminant Color Estimation by Hue Categorization Based on Gray World Assumption

    Harumi Kawamura, Shunichi Yonemura, Jun Ohya, Norihiko Matsuura

    COMPUTATIONAL IMAGING IX   7873 ( 787312 ) 1 - 12  2011  [Refereed]

     View Summary

    This paper proposes a gray world assumption based method for estimating an illuminant color from an image by hue categorization. The gray world assumption hypothesizes that the average color of all the objects in a scene is gray. However, it is difficult to estimate an illuminant color correctly if the colors of the objects in a scene are dominated by certain colors. To solve this problem, our method uses the opponent color properties that the average of a pair of opponent colors is gray. Thus our method roughly categorizes the colors derived from the image based on hue and selects them one by one from the hue categories until selected colors satisfy the gray world assumption. In our experiments, we used three kinds of illuminants (i.e., CIE standard illuminants A and D-65, and a fluorescent light) and two kinds of data sets. One data set satisfies the gray world assumption, and the other does not. Experiment results show that estimated illuminants are closer to the correct ones than those obtained with the conventional method and the estimation error for both using CIE standard illuminants A and D-65 by our method are within the barely noticeable difference in human color perception.

    DOI

  • OMR of Early Plainchant Manuscripts in Square Notation: a Two-stage System

    Carolina Ramirez, Jun Ohya

    DOCUMENT RECOGNITION AND RETRIEVAL XVIII   7874 ( 787417 ) 1 - 8  2011  [Refereed]

     View Summary

    While Optical Music Recognition (OMR) of modern printed and handwritten documents is considered a solved problem, with many commercial systems available today, the OMR of ancient musical manuscripts still remains an open problem. In this paper we present a system for the OMR of degraded western plainchant manuscripts in square notation from the XIV to XVI centuries. The system has two main blocks, the first one deals with symbol extraction and recognition, while the second one acts as an error detection stage for the first block outputs. For symbol extraction we use widely known image-processing techniques, such as Sobel filtering and Hough Transform, and SVM for classification. The error detection stage is implemented with a hidden Markov model (HMM), which takes advantage of a priori knowledge for this specific kind of music.

    DOI

  • Appearance-based Human Gesture Recognition using Multimodal Features for Human Computer Interaction

    Dan Luo, Hua Gao, Hazim Kemal Ekenel, Jun Ohya

    HUMAN VISION AND ELECTRONIC IMAGING XVI   7865 ( 786509 ) 1 - 6  2011  [Refereed]

     View Summary

    The use of gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and LDA is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.

    DOI

  • Study of Recognizing Human Motion Observed from an Arbitrary Viewpoint Based on Decomposition of a Tensor Containing Multiple View Motions

    Takayuki Hori, Jun Ohya, Jun Kurumisawa

    COMPUTATIONAL IMAGING IX   7873 ( 787310 ) 1 - 10  2011  [Refereed]

     View Summary

    We propose a Tensor Decomposition based algorithm that recognizes the observed action performed by an unknown person and unknown viewpoint not included in the database. Our previous research aimed motion recognition from one single viewpoint. In this paper, we extend our approach for human motion recognition from an arbitrary viewpoint. To achieve this issue, we set tensor database which are multi-dimensional vectors with dimensions corresponding to human models, viewpoint angles, and action classes. The value of a tensor for a given combination of human silhouette model, viewpoint angle, and action class is the series of mesh feature vectors calculated each frame sequence. To recognize human motion, the actions of one of the persons in the tensor are replaced by the synthesized actions. Then, the core tensor for the replaced tensor is computed. This process is repeated for each combination of action, person, and viewpoint. For each iteration, the difference between the replaced and original core tensors is computed. The assumption that gives the minimal difference is the action recognition result. The recognition results show the validity of our proposed method, the method is experimentally compared with Nearest Neighbor rule. Our proposed method is very stable as each action was recognized with over 75% accuracy.

    DOI

  • Study of Temporal Modified-RANSAC Based Method for the Extraction and 3D Shape Reconstruction of Moving Objects from Dynamic Stereo Images and for Estimating the Camera Pose

    Naotomo Tatematsu, Jun Ohya

    INTELLIGENT ROBOTS AND COMPUTER VISION XXVIII: ALGORITHMS AND TECHNIQUES   7878 ( 78780E ) 1 - 13  2011  [Refereed]

     View Summary

    This paper proposes a Temporal Modified-RANSAC based method that can discriminate each moving object from the still background in the stereo video sequences acquired by moving stereo cameras, can compute the stereo cameras' egomotion, and can reconstruct the 3D structure of each moving object and the background. We compute 3D optical flows from the depth map and results of tracking feature points. We define "3D flow region" as a set of connected pixels whose 3D optical flows have a common rotation matrix and translation vector. Our Temporal Modified-RANSAC segments the detected 3D optical flows into 3D flow regions and computes the rotation matrix and translation vector for each 3D flow region. As opposed to the conventional Modified-RANSAC for only two frames, The Temporal Modified-RANSAC can handle temporal images with arbitrary length by performing the Modified-RANSAC to the set of a 3D flow region that classified in the latest frame and new 3D optical flows detected in the current frame iteratively. Finally, the 3D points computed from the depth map in all the frames are registered using each 3D flow region's matrix to the initial positions in the initial frame so that the 3D structures of the moving objects and still background are reconstructed. Experiments using multiple moving objects and real stereo sequences demonstrate promising results of our proposed method.

    DOI

  • Efficient, High-Quality, GPU-Based Visualization of Voxelized Surface Data with Fine and Complicated Structures

    Sven Forstmann, Jun Ohya

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E93D ( 11 ) 3088 - 3099  2010.11  [Refereed]

     View Summary

    This paper proposes a GPU based method that can visualize voxelized surface data with fine and complicated features has high rendering quality at interactive frame rates and provides low memory consumption The surface data is compressed using run length encoding (RLE) for each level of detail (LOD) Then the loop for the rendering process is performed on the GPU for the position of the viewpoint at each time instant The scene is raycasted in planes where each plane is perpendicular to the horizontal plane in the world coordinate system and passes through the viewpoint For each plane one ray is cast to rasterize all RLE elements intersecting this plane starting from the viewpoint and ranging up to the maximum view distance This rasteritization process projects each RLE element passing the occlusion test onto the screen at a LOD that decreases with the distance of the RLE element from the viewpoint Finally the smoothing of voxels in screen space and full screen anti aliasing is performed To provide lighting calculations without storing the normal vector inside the RLE data structure our algorithm recovers the normal vectors from the rendered scene s depth butter After the viewpoint changes the same process is re executed for the new viewpoint Experiments using different scenes have shown that the proposed algorithm is faster than the equivalent CPU implementation and other related methods Our experiments further prove that this method is memory efficient and achieves high quality results.

    DOI

  • Studies on Enterprises’ Adaptive Strategy to Economic Responsibility under EPR

    Yiping Zhao, Jun Ohya

    Proceedings of 2nd International Conference on Sustainable Enterprises of the Future     254 - 257  2010.09  [Refereed]

  • 色相に着目した灰色仮説に基づく照明光推定法の一提案

    川村春美, 米村俊一, 大谷淳, 松浦宣彦

    情報処理学会研究報告   2010-AVM-70 ( 6 ) 1 - 6  2010.08

    J-GLOBAL

  • Symbol Classification Approach for OMR of Square Notation Manuscripts

    Carolina Ramirez, Jun Ohya

    Proceedings of the 11th International Society for Music Information Retrieval Conference (ISMIR 2010)     549 - 554  2010.08  [Refereed]

  • Study on Adaptive Management Model and Key Factors under Extended Producer Responsibility

    Yiping Zhao, Jun Ohya

    GITS/GITI 紀要 2009-2010 (早稲田大学大学院国際情報通信研究科 国際情報通信研究センター), 査読論文     136 - 144  2010.07  [Refereed]

  • Elliptical Object Detection by a Modified RANSAC with Sampling Constraint from Boundary Curves' Clustering

    Yingdi Xie, Jun Ohya

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E93D ( 3 ) 611 - 623  2010.03  [Refereed]

     View Summary

    This paper proposes a method for detecting ellipses from an image despite (1) multiple colors within the ellipses, (2) partially occluded ellipses' boundaries, (3) noisy. locally deformed boundaries of ellipses, (4) presence of multiple objects other than the ellipses in the image, and (5) combinations of (1) through (4). After boundary curves are obtained by edge detection, by utilizing the first-order difference curves of the edge orientation of each pixel in the boundary curves, a segment-reconnect method obtains boundary clusters. Then, a modified RANSAC detects ellipses by choosing five pixels randomly from the boundary clusters, where overlapped ellipses are merged. Experimental results using synthesized images and real images demonstrate the effectiveness of the proposed method together with comparison with the Randomized Hough Transform, a well-known conventional method.

    DOI

  • Study of Recognizing Multiple Persons' Complicated Hand Gestures from the Video Sequence Acquired by a Moving Camera

    Luo Dan, Jun Ohya

    HUMAN VISION AND ELECTRONIC IMAGING XV   7527  2010  [Refereed]

     View Summary

    Recognizing hand gestures from the video sequence acquired by a dynamic camera could be a useful interface between humans and mobile robots. We develop a state based approach to extract and recognize hand gestures from moving camera images. We improved Human-Following Local Coordinate (HFLC) System, a very simple and stable method for extracting hand motion trajectories, which is obtained from the located human face, body part and hand blob changing factor. Condensation algorithm and PCA-based algorithm was performed to recognize extracted hand trajectories. In last research, this Condensation Algorithm based method only applied for one person's hand gestures. In this paper, we propose a principal component analysis (PCA) based approach to improve the recognition accuracy. For further improvement, temporal changes in the observed hand area changing factor are utilized as new image features to be stored in the database after being analyzed by PCA. Every hand gesture trajectory in the database is classified into either one hand gesture categories, two hand gesture categories, or temporal changes in hand blob changes. We demonstrate the effectiveness of the proposed method by conducting experiments on 45 kinds of sign language based Japanese and American Sign Language gestures obtained from 5 people. Our experimental recognition results show better performance is obtained by PCA based approach than the Condensation algorithm based method.

    DOI

  • Study on Incentive Management of the End-of-life Products Collection and Recycling under EPR

    Yiping Zhao, Jun Ohya

    PROCEEDINGS OF 2010 INTERNATIONAL CONFERENCE ON ENVIRONMENTAL SCIENCE AND DEVELOPMENT     157 - 162  2010  [Refereed]

     View Summary

    In this paper, the end-of-life products (ELPs) collecting and recycling system is studied under the application of extended producer responsibility (EPR) programs. Economic performance is evaluated based on the decomposition and analysis of major cost and revenues components during the recycling process. We identified two problematic scenarios where EPR programs cannot run smoothly. The according prosperous incentive management strategies are discussed as the corresponding solution.

  • Integrated System for End-of-life Products Collecting and Recycling under Extended Producer Responsibility

    Yiping Zhao, Jun Ohya

    The Sixth International Conference on Technology, Knowledge and Society, Berlin     1 - 7  2010.01  [Refereed]

  • PCA-based Method for Recognizing Multiple Persons’ Hand Gestures from the Video Sequence Acquired by a Moving Camera

    Luo Dan, Jun Ohya

    SPIE-IS&T, Electronic Imaging2010   7527   75271N  2010.01  [Refereed]

  • Identifying a Walking Human by a Tensor Decomposition Based Approach and Tracking the Human across Discontinuous Fields of Views of Multiple Cameras

    Takayuki Hori, Jun Ohya, Jun Kurumisawa

    COMPUTATIONAL IMAGING VIII   7533   75330X  2010  [Refereed]

     View Summary

    This paper proposes a method that identifies and tracks a walking human across discontinuous fields of views of multiple cameras for the purpose of video surveillance. A typical video surveillance system has multiple cameras, but there are several spaces within the surveillance area that are not within any of the camera's field of view. Also, there are discontinuities between the fields of views of adjacent cameras. In such a system, humans need to be tracked across discontinuous fields of views of multiple cameras. Our proposed model addresses this issue using the concepts of gait pattern, gait model, and motion signature. Each human's gait pattern is constructed and stored in a database. This gait pattern spans a tensor space that consists of three dimensions: person, image feature, and spatio-temporal data. A human's gait model can be constructed from the gait pattern using the "tensor decomposition based approach" described in this paper. When human(s) appears in one of the camera's field of a view (which is often discontinuous from the other camera's field of views), the human's motion signature is calculated and compared to each person in the database's gait model. The person with the gait model that is most similar to the motion signature is identified as same person. After the person is identified, the person is tracked within the field of view of the camera using the mean-shift algorithm based on color parameters. We conducted two experiments; the first experiment was identifying and tracking humans in a single video sequence, and experiments, the percentage of subjects that were correctly identified and tracked was better than that of two currently widely-used methods, PCA and nearest-neighbor. In the second experiment was the same as the first experiment but consisted of multiple-cameras with discontinuous views. The second experiment (human tracking across discontinuous images), shows the potential validity of the proposed method in a typical surveillance system.

    DOI

  • A Method for Detecting Multiple Independently Moving Objects from the Sequences Acquired by Active Stereo Cameras and Estimating the Cameras' Egomotion

    Yingdi Xie, Jun Ohya

    Journal of the Institute of Image Electronics Engineers of Japan   39 ( 2 ) 163 - 174  2010  [Refereed]

     View Summary

    This paper proposes a modified RANSAC based method that can detect multiple independently moving objects from the video sequences acquired by active stereo cameras, and can estimate the egomotion of the stereo cameras. We first compute 3D optical flows from consecutive frames, where dense 3D optical flows, which are needed for the subsequent egomotion estimation based on the modified RANSAC, are obtained by the process proposed in this paper. From the dense flows, three 3D optical flows are randomly selected for computing an egomotion (model). The validity of that egomotion model is checked using the weight obtained by calculating a consistency between that model and each of the dense flows. If it turns out that the estimated egomotion model is valid, the label that identifies this model is assigned to the inliers, which are consistent with this model, and then the labeled inliers are deleted. The random choice of three flows and its subsequent processes are repeated till one of the termination con itions is met. Using the inliers present in the static background, the real egomotion is computed. Experiments using synthesized and real stereo sequences demonstrate the effectiveness of the proposed method. © 2010, The Institute of Image Electronics Engineers of Japan. All rights reserved.

    DOI

  • An improved Kernel-based fuzzy C-means algorithm with spatial information for brain MR image segmentation

    Rong Xu, Jun Ohya

    International Conference Image and Vision Computing New Zealand   173   1 - 7  2010  [Refereed]

     View Summary

    In this paper, we propose an improved Kernel-based Fuzzy C-means Algorithm (iKFCM) with spatial information to reduce the effect of noise for brain MR image segmentation. We use k-nearest neighbour model and a neighbourhood controlling factor by estimating image contextual constraints to optimize the objective function of conventional KFCM method. Conventional KFCM algorithms classify each pixel in image only by its own gray value, but the proposed method classifies by the gray values of its neighbourhood system. For this reason, the proposed iKFCM has a strong robustness for image noise in image segmentation. In experiments, some synthetic grayscale images and simulated brain MR images are used to assess the performance of iKFCM in comparison with other fuzzy clustering methods. The experimental results show that the proposed iKFCM method achieves a better segmentation performance than other fuzzy clustering methods. © 2010 IEEE.

    DOI

  • Feasibility of using machine vision based cash register system in a University Cafeteria

    Khanh N. Do, Jun Ohya, Davar Pishva, G. G D Nishantha

    Proceedings - 2010 3rd International Congress on Image and Signal Processing, CISP 2010   6   2554 - 2558  2010  [Refereed]

     View Summary

    This paper explores the feasibility of using machined vision based cash register system in a university cafeteria and automatically determine price of the food items that are contained on a cafeteria food tray. The idea is to take an image of the food tray, identify all the food items on the tray, and display the total food cost. A fully automated version of the system is envisioned to use a standard biometric technology for customer identification and for automatic debit process of the food cost from the customer's pre-registered account. The analysis method uses shape and size analysis to identify food items that are served on distinct type dishes and color analysis for those items that are served in similar type dishes. The paper explains prototype of a developed system and experimental results for 10 types of meals served at Ritsumeikan Asia Pacific University Cafeteria to justify the validity of this approach. ©2010 IEEE.

    DOI

  • Application of extended producer responsibility in Chinese WEEE management

    Yiping Zhao, Jun Ohya

    ICCCE 2010 - 2010 International Conference on Chemistry and Chemical Engineering, Proceedings     272 - 275  2010  [Refereed]

     View Summary

    In this paper, the administration-driven WEEE management system in Europe and the market-driven end-of-life automobile management system in the US are studied first in this paper. Successful experiences and deficiencies of each system are investigated and compared. Then the potentials for integrating these two systems under each single-instrument-dominant scenario are identified followed by the discussion on key factors for soundly building an integrated system under various scenarios in general. Based on these achievements, the WEEE management system in China is taken as an example to demonstrate how to practically set up an integrated system in such a specific conditions, so that the application of EPR principle could be better promoted in a more economic efficient and ecological effective manner. © 2010 IEEE.

    DOI

  • STUDY ON HUMAN GESTURE RECOGNITION FROM MOVING CAMERA IMAGES

    Dan Luo, Jun Ohya

    2010 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME 2010)     274 - 279  2010  [Refereed]

     View Summary

    We develop a framework based approach to extract and recognize hand gestures from the video sequence acquired by a dynamic camera, which could be a useful interface between humans and mobile robots. We use Human-Following Local Coordinate (HFLC) System, a very simple and stable method for extracting hand motion trajectories, which is obtained from the located human face and body part. Hand trajectory motion models (HTMM) are constructed by HFLC and hand blob changing factor. In this paper, we apply a principal component analysis (PCA) based approach to improve the recognition accuracy. For further improvement, temporal changes in the observed hand area changing factor are utilized as new image features to be stored in the database after being analyzed by PCA. Each HTMM in the database is classified into gesture categories, or temporal changes in hand blob changes. We demonstrate the effectiveness of the proposed method by conducting experiments on 51 kinds of sign language based Japanese and American Sign Language gestures obtained from 7 people. Our experimental recognition results show better performance is obtained by PCA based approach than the Condensation algorithm based method.

    DOI

  • Visual Support System Considering Privacy Protection by Using the Line Drawing Video Image

    Shun-Ichi Yonemurat, Ken Tsutsuguchit, Satoshi Shimadat, Akira Suzukit, Yukio Tokunaga, Jun Ohya

    Journal of the Institute of Image Electronics Engineers of Japan   39 ( 4 ) 481 - 489  2010  [Refereed]

     View Summary

    This paper proposes a system that automatically deforms the original video images into line drawing expression transmitted in both directions so as to achieve two goals: easing the concerns of users privacy and ensuring good enough support by visual channel. We examine the effect of line drawing video for support system on the effectiveness and impression of privacy on a collaboration between a novice user and an operator. As a result, it became clear that concern of the user on privacy or security are greatly reduced by using a line drawing image, and there was no significant difference on efficiency with an original image and a line drawing image. © 2010, The Institute of Image Electronics Engineers of Japan. All rights reserved.

    DOI

  • •” Media: Text Based CMC Tool Which Touch off Informal Communication

    Shun-Ichi Yonemura, Satoshi Shimada, Chen Li Jen, Jun Ohya, Yukio Tokunaga

    Journal of the Institute of Image Electronics Engineers of Japan   39 ( 4 ) 442 - 450  2010  [Refereed]

     View Summary

    This paper describes about the media which touch off informal text communication by black spot stimulus. A black spot stimulus is shown as a background image of the text field in advance of text input. The purpose of black spot stimulus presentation is to touch off conceiving topic based on the cognitive process model of creative thinking. The prototype system based on this media concept was developed, and the communication experiment was conducted. It turned out as a result of the experiment that these media touched off the idea about making topic strongly in the text communication. Moreover, it was implied that the topic touched off deepens the mutual understanding between speakers. © 2010, The Institute of Image Electronics Engineers of Japan. All rights reserved.

    DOI

  • 中国麦積山石窟のデジタルアーカイブとVRコンテンツ

    阿部信明, 河合隆史, 大谷 淳, 査 紅彬, 安藤 真

    日本バーチャルリアリティ学会論文誌   14 ( 3 ) 275 - 282  2009.09  [Refereed]

    DOI

  • Elliptical Shaped Object Recognition via a Modified RANSAC with Edge Orientation Curve’s Segmentation-Merge

    Yingdi Xie, Jun Ohya

    Proc. Visualization, Imaging, and Image Processing   652-028   33 - 40  2009.07  [Refereed]

  • Company-Level Adaptive Management Model under Extended Producer Responsibility

    Yiping Zhao, Yingdi Xie, Jun Ohya

    Proceeding of The 24th International Conference on Solid Waste Technology and Management   24   804 - 813  2009.03  [Refereed]

  • Efficient Detection of Ellipses from an Image by a Guided Modified RANSAC

    Yingdi Xie, Jun Ohya

    IMAGE PROCESSING: ALGORITHMS AND SYSTEMS VII   7245   72450W  2009  [Refereed]

     View Summary

    In this paper, we propose a novel ellipse detection method which is based on a modified RANSAC, with automatic sampling guidance from the edge orientation difference curve. Hough Transform family is one of the most popular and methods for shape detection, but the Standard Hough Transform loses its computation efficiency if the dimension of the parameter space gets high. Randomized Hough Transform, an improved version of Standard Hough Transform has difficulty in detecting shapes from complicated, cluttered scenes because of its random sampling process. As a pre-process for random selection of five pixels to be used to build the ellipse's equation, we propose a two-step algorithm: (1) region segmentation and contour detection by mean shift algorithm (2) contour splitting based on the edge orientation difference curve obtained from the contour of each region. In each contour segment obtained by step (2), 5 pixels are randomly selected and the modified RANSAC is applied to the 5 pixels so that an accurate ellipse model is obtained. Experimental result show that the proposed method can achieve high accuracies and low computation cost in detecting multiple ellipses from an image.

    DOI

  • Comparative Study of Methods for Recognition an Unknown Person's Action from a Video Sequence

    Takayuki Hori, Jun Ohya, Jun Kurumisawa

    IMAGE PROCESSING: ALGORITHMS AND SYSTEMS VII   7245   72450V  2009  [Refereed]

     View Summary

    This paper proposes a Tensor Decomposition Based method that can recognize an unknown person's action from a video sequence, where the unknown person is not included in the database (tensor) used for the recognition. The tensor consists of persons, actions and time-series image features. For the observed unknown person's action, one of the actions stored in the tensor is assumed. Using the motion signature obtained from the assumption, the unknown person's actions are synthesized. The actions of one of the persons in the tensor are replaced by the synthesized actions. Then, the core tensor for the replaced tensor is computed. This process is repeated for the actions and persons. For each iteration, the difference between the replaced and original core tensors is computed. The assumption that gives the minimal difference is the action recognition result. For the time-series image features to be stored in the tensor and to be extracted from the observed video sequence, the human body silhouette's contour shape based feature is used. To show the validity of our proposed method, our proposed method is experimentally compared with Nearest Neighbor rule and Principal Component analysis based method. Experiments using 33 persons' seven kinds of action show that our proposed method achieves better recognition accuracies for the seven actions than the other methods.

    DOI

  • Hand-gesture extraction and recognition from the video sequence acquired by a dynamic camera using condensation algorithm

    Luo Dan, Jun Ohya

    Proceedings of SPIE - The International Society for Optical Engineering   7252   72520S  2009  [Refereed]

     View Summary

    To achieve environments in which humans and mobile robots co-exist, technologies for recognizing hand gestures from the video sequence acquired by a dynamic camera could be useful for human-to-robot interface systems. Most of conventional hand gesture technologies deal with only still camera images. This paper proposes a very simple and stable method for extracting hand motion trajectories based on the Human-Following Local Coordinate System (HFLC System), which is obtained from the located human face and both hands. Then, we apply Condensation Algorithm to the extracted hand trajectories so that the hand motion is recognized. We demonstrate the effectiveness of the proposed method by conducting experiments on 35 kinds of sign language based hand gestures. © 2009 SPIE-IS&amp
    T.

    DOI

  • Prompter "center dot" Based Creating Thinking Support Communication System That Allows Hand-Drawing

    Li Jen Chen, Jun Ohya, Shunichi Yonemura, Sven Forstmann, Yukio Tokunaga

    HUMAN-COMPUTER INTERACTION, PT II   5611   783 - +  2009  [Refereed]

     View Summary

    Research into creative thinking-support tools and communication is commonly focused on how to develop and share ideas between participants or with others. In this paper, we proposes a creative thinking support method that utilizes randomly generated visual prompter (black circle) image patterns (VP-patterns) and free hand-drawing and writing functions. Concepts and ideas of the research have been explained together with the development of the systems (CSP1 and CSP2). Experiments have been Conducted in order to evaluate the potentials and effectiveness of the system. From the results, a tendency towards inspiring creative ideas by participants has been observed.

    DOI

  • 3D Sensation Enhancement of 3D Digitized Models for 2D Displays Using Local Gray-level Contrast and Range Information

    Xiao Ding, Jun Ohya

    2008KJPR(Third Korea-Japan Joint Workshop on Pattern Recognition)     55 - 56  2008.11  [Refereed]

  • Contrast Enhancement of Color Images based on PCA of Luminance and Saturation

    Xiao Ding, Jun Ohya

    2008KJPR(Third Korea-Japan Joint Workshop on Pattern Recognition)     9 - 10  2008.11  [Refereed]

  • Development and Evaluation of the System for Educational Method Using Students’ Understanding Degree

    Takaaki Ohkawauchi, Jun Ohya, Yoshikazu Tateno

    ICoME 2008“Enhancing ICT in Education for Global Society”     266 - 273  2008.08  [Refereed]

  • Body-part Segmentation for a Real Human Wearing Multiple-Colored Suit by a Mean Shift Based Algorithm

    Huimin Guo, Jun Okazaki, Dong-Wan Kang, Jun Ohya

    2007KJPR (Second Korea-Japan Joint Workshop)    2007.10  [Refereed]

  • Accurate, Real-time 3D Reconstruction of Human Postures from Orthogonal Images Based on Neural Networks

    Dong-Wan Kang, Zhongze Chen, Jun Ohya

    2007KJPR (Second Korea-Japan Joint Workshop)     111 - 116  2007.10  [Refereed]

  • ACCESSIBLE INFORMATION PRESENTATION METHOD IN PUBLIC SPACE

    Shunichi Yonemura Shinichiro Eitoku Kenichiro Shimokura, Jun Ohya

    Proceedings of Workshop on Tactic and Haptic Interaction     70 - 75  2007.05  [Refereed]

  • MULTIMODAL COMMUNICATION ON VISUAL SUPPORT SYSTEM

    Shunichi Yonemura, Tohru Yoshida, Yukio Tokunaga, Jun Ohya

    Proceedings of Workshop on Tactic and Haptic Interaction     64 - 69  2007.05  [Refereed]

  • A STUDY OF A COMPUTER MEDIATED COMMUNICATION VIA THE “●” PROMPT SYSTEM - Introducting the affects on the stimulation of thought processes and the inspiration of creative ideas -,

    Li Jen Chen, Jun Ohya, Shunichi Yonemura, Nobuyuki, Harada,Yukio Tokunaga

    Proceedings of Workshop on Tactic and Haptic Interaction     50 - 55  2007.05  [Refereed]

  • 逆S字型変換による濃淡画像のコントラスト強調手法

    小林裕一, 加藤俊一, 大谷 淳

    電子情報通信学会論文誌 D   Vol. J90-D ( No. 5 ) 1263 - 1274  2007.05  [Refereed]

  • 境界要素法に基づく3次元多関節弾性物体の効率的かつ頑健な挙動シミュレーション

    渡辺隆史, 大谷 淳

    日本バーチャルリアリティ学会論文誌   12 ( 1 ) 85 - 94  2007.03  [Refereed]

    DOI

  • Efficient multiple independent motion segmentation from an active platform by utilizing modified RANSAC

    Yingdi Xie, Jun Ohya

    PROCEEDINGS OF THE IASTED INTERNATIONAL CONFERENCE ON GRAPHICS AND VISUALIZATION IN ENGINEERING     31 - +  2007  [Refereed]

     View Summary

    hi this paper, an efficient approach to segmentation of different independent motion areas from a moving platform is described. This approach is implemented on a stereo vision system, depth information could be computed by matching feature points between stereo images. For two consecutive frames, ego-motion is estimated from the optical flows, including depth information belonging to the background, which has a larger space distribution comparing to those of independent moving objects. In order to distinguish different motion areas, we proposed a modified version of RANSAC mechanism, which could handle the problem of multiple model extraction in a noisy environment.

  • Contrast enhancement of color images based on wavelet transform and human visual system

    Ding Xiao, Jun Ohya

    PROCEEDINGS OF THE IASTED INTERNATIONAL CONFERENCE ON GRAPHICS AND VISUALIZATION IN ENGINEERING     58 - +  2007  [Refereed]

     View Summary

    This paper proposes a new method for enhancing the contrast of color images based on Wavelet Transform and human visual system. The RGB (red, green, and blue) values of each pixel in a color image are converted to HSV (hue, saturation and value) values. To the V (luminance value) components of the color image, Wavelet Transform is applied so that the V components are decomposed into the approximate components and detail components. The obtained coefficients of the approximate components are converted by a grey-level contrast enhancement technique based on human visual system. Then, inverse Wavelet transform is performed for the converted coefficients so that the enhanced V values are obtained. The S components are enhanced by histogram equalization. The H components are not changed, because changes in the H components could degrade the color balance between the HSV components. The enhanced S and V together with H are converted back to RGB values. The effectiveness of the proposed method is demonstrated experimentally.

  • Efficient, robust, bem (boundary element method) based method for simulating articulated 3d elastic object models

    Takafumi Watanabe, Jun Ohya

    PROCEEDINGS OF THE IASTED INTERNATIONAL CONFERENCE ON GRAPHICS AND VISUALIZATION IN ENGINEERING     7 - +  2007  [Refereed]

     View Summary

    This paper proposes an efficient and effective computation method for simulating articulated 3D elastic objects. The articulated elastic object consists of bones, elastic object primitives and joining surfaces. After transforming the bones by inverse kinematics, the elastic object primitives are deformed basically by the BEM (Boundary Element Method). To reduce the computation cost and memory usage and to achieve robust deformations, this paper improves the BEM based module of the conventional method: more specifically, a new method for computing the displacements of the vertices in the joining surfaces. Experimental results show that the proposed model can be simulated 10 times as fast as the conventional model or faster, and that real-time performances such as 60fps can be achieved even for models with 20,000 vertices. High quality simulations of different deformations including expansions and contractions are achieved using character type models.

  • Deformation Styles for Spline-based Skeletal Animation

    Sven Forstmann, Jun Ohya, Artus Krohn-Grimberghe, Ryan McDougall

    SYMPOSIUM ON COMPUTER ANIMATION 2007: ACM SIGGRAPH/ EUROGRAPHICS SYMPOSIUM PROCEEDINGS     141 - 150  2007  [Refereed]

     View Summary

    We present a novel skinned skeletal animation system based on spline-aligned deformations for providing high quality and fully designable deformations in real-time. Our ambition is to allow artists the easy creation of abstract, pose-dependent deformation behaviors that might directly be assigned to a large variety of target objects simultaneously To achieve this goal, we introduce the usage of deformation styles and demonstrate their applicability by our animation system. We therefore enhance spline-skinned skeletal animation with two sweep-based free-form-deformation (FFD) variants. The two FFD variants are pose-dependent, driven by three textures and three curves, which can be designed by the artist. As the three textures are similar to height-maps, their creation is very intuitive. Once designed, the deformation styles can be directly applied to any number of targets for imitating material behaviors of cloth, metal or even muscles. Our GPU based implementation shows promising results for real-time usage, as about 30 Million vertices per second can be animated. The basic spline-skinning even reaches more than twice the speed and gets close to the performance of skeletal subspace deformation (SSD). Furthermore, our method can easily be combined along with other existing deformation techniques as pose space deformation or SSD.

  • Communication analysis of visual support system that uses line drawing expression

    Shunichi Yonemura, Tohru Yoshida, Yukio Tokunaga, Jun Ohya

    HUMAN INTERFACE AND THE MANAGEMENT OF INFORMATION: METHODS, TECHNIQUES AND TOOLS IN INFORMATION DESIGN, PT 1, PROCEEDINGS   4557   359 - 365  2007  [Refereed]

     View Summary

    This paper proposes a system that automatically deforms the images transmitted in both directions so as to achieve two goals: easing the concerns of users and ensuring well enough support. We examine here the effect of visual information quality on the effectiveness and impression of collaboration between a novice user and an operator.

  • A study of a computer mediated communication via the "circle" prompt system - Analysis of the affects on the stimulation of thought processes and the inspiration of creative ideas

    Li Jen Chen, Nobuyuki Harada, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    2007 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-5     1119 - +  2007  [Refereed]

     View Summary

    Research into thinking-support tools is commonly focused on how to develop and share ideas between participants or with others. In this paper, we propose and develop a communication system that stimulates the thought processes and inspires the creative ideas of participants by using a visual "circle" prompt within the framework of a communication pallet. Experiments have been conducted into methods of stimulating the thought process and inspiring ideas during conversation and the results have been analyzed. From the results, a tendency towards inspiring creative ideas by participants has been observed.

  • A study of a computer mediated communication via the "circle" prompt system - Analysis of the affects on the stimulation of thought processes and the inspiration of creative ideas

    Li Jen Chen, Nobuyuki Harada, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    2007 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-5     1119 - +  2007  [Refereed]

     View Summary

    Research into thinking-support tools is commonly focused on how to develop and share ideas between participants or with others. In this paper, we propose and develop a communication system that stimulates the thought processes and inspires the creative ideas of participants by using a visual "circle" prompt within the framework of a communication pallet. Experiments have been conducted into methods of stimulating the thought process and inspiring ideas during conversation and the results have been analyzed. From the results, a tendency towards inspiring creative ideas by participants has been observed.

  • 伸縮を含む対話的変形操作が可能な多関節三次元弾性物体モデル

    渡辺隆史, 大谷 淳, 楜沢 順, 徳永幸生

    電子情報通信学会論文誌D   J89-D ( 10 ) 2145 - 2154  2006.10  [Refereed]

  • やわらかさ・かたさの判別に係る視覚的要因に関する一考察

    小林裕一, 大谷 淳

    GITS/GITI紀要、早稲田大学大学院国際情報通信研究科/国際情報通信研究センター     178 - 187  2006.07  [Refereed]

  • VISUAL BLENDING USING 2D MORPHING

    S.K. Semwal, J. Ohya, I. Semwal, B Bikker

    Proceedings of the 24th IASTED International Multi-Conference Signal Processing, Pattern Recognition, and Applications     316 - 321  2006.02  [Refereed]

  • Fast Skeletal Animation by skinned Arc-Spline based Deformation

    Sven Forstmann, Jun Ohya

    EUROGRAPHICS 2006: SHORT PAPERS     1 - 4  2006  [Refereed]

     View Summary

    Presented is a novel skeletal animation system for providing high quality geometric deformations in real-time. Each bone of the skeleton is therefore represented by a spline, rather than using conventional matrix rotation. In our approach, each vertex of the animated character can be influenced by a maximum of three spline-curves, which is sufficient for skinned animation. One spline is parameterized by three control points and a possible twist. As opposed to conventional Bezier curves does our arc-spline rely on trigonometric functions for providing better curvatures. The optimized implementation using the OpenGL-shading language shows very promising results for real-time character animation, as even about 1 Million vertices were able to be transformed at interactive 43 frames per second on a GeForce 7800 GTX graphics card.

  • EM-in-M: Analyze and synthesize emotion in motion

    Yuichi Kobayashi, Jun Ohya

    ADVANCES IN MACHINE VISION, IMAGE PROCESSING, AND PATTERN ANALYSIS   4153   135 - 143  2006  [Refereed]

     View Summary

    We have been researching the relationship between human motion and emotion. In this paper, our purpose is to extract motion features specific to each emotion. We propose a new approach for motion data analysis, which applies the higher order Singular Value Decomposition(HOSVD) direct to the motion data and the wavelet analysis to the synthesized data with SVD. The HOSVD models the mapping between persons and emotions. The model can synthesize a complete data acting with each emotion for a given new person. The wavelet analysis extracts each motion feature from the synthesized data for each emotion. Some experimental results using motion capture data for "gait" action and 6 emotions - "angry, joy, sad and so on" show that our method can synthesize novel gait motions for a person by using the extracted motion elements and can extract some features specific to each emotion.

  • 2段階境界要素法を用いる三次元弾性物体の変形と移動の実時間アニメーション法

    渡辺隆史, 大谷 淳, 楜沢 順, 徳永幸生

    電子情報通信学会論文誌 D-Ⅱ   J88-D-Ⅱ ( 9 ) 1876 - 1888  2005.09  [Refereed]

  • Visualization of large iso-surfaces based on nested clip-boxes

    Sven Forstmann, Jun Ohya

    ACM SIGGRAPH 2005 Posters, SIGGRAPH 2005   Conference Select CD-ROM Disc   126  2005.07  [Refereed]

    DOI

  • 顔面筋動作に基づく形状変形とテクスチャの張替えによる三次元人物顔画像の再構築

    佐藤一, 大谷 淳, 寺島信義

    電子情報通信学会論文誌 D-Ⅱ   J88-D-Ⅱ ( 7 ) 1126 - 1142  2005.07  [Refereed]

  • Visual-dimension interact system (VIS)

    A Onda, T Oku, E Yu, Y Lee, Choh, I, PY Chiu, J Ohya

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E88D ( 5 ) 947 - 953  2005.05  [Refereed]

     View Summary

    In this paper we describe a mixed reality-supported interactive viewing enhancement museum display system: Visual-dimension Interact System (VIS). With a transparent interactive inter-face, the museum visitor is able to see, manipulate, and interact with the physical exhibit and its virtual information, which are overlapped on one other. Furthermore, this system provides the possibility for visitor to experience the creation process in an environment as close as possible to the real process. This has the function of assisting the viewer in understanding the exhibit and most importantly, gaining a so-to-speak hands-on experience of the creation process itself leading to a deeper understanding of it.

    DOI

  • Analysis of expressing audiences in a cyber-theater

    K DongWan, K Huang, J Ohya

    2005 IEEE International Conference on Multimedia and Expo (ICME), Vols 1 and 2   CD-ROM Proceedings, 4 pages   727 - 730  2005  [Refereed]

     View Summary

    This paper studies how audiences should be expressed in a Cyber-theater, in which remotely located persons can direct plays as directors, perform as performers and/or see the performances as audiences through a networked virtual environment. It is noted that the audience effect has been widely acknowledged in the real-world theater: that is, the audience reaction has a significant effect on the acting of player and performance of the play itself. However, only a few works relevant to audiences in the Cyber Theater can be seen. This paper studies whether the audience effect exists also in the Cyber-theater. By constructing a system in which two actors are displayed a remotely located audience's avatar in which the audience can display his/her emotional actions, we clarified that interactions between the actors and audience are effective.

    DOI

  • A study of synthesizing new human motions from sampled motions using tensor decomposition

    R Kalanov, J Cho, J Ohya

    2005 IEEE International Conference on Multimedia and Expo (ICME), Vols 1 and 2   CD-ROM Proceedings, 4 pages   1327 - 1330  2005  [Refereed]

     View Summary

    This paper applies an algorithm, based on Tensor Decomposition, to a new synthesis application: by using sampled motions of people of different ages under different emotional states, new motions for other people are synthesized Human motion is the composite consequence of multiple elements, including the action performed and a motion signature that captures the distinctive pattern of movement of a particular individual. By performing decomposition, based on N-mode SVD (singular value decomposition), the algorithm analyzes motion data spanning multiple subjects performing different actions to extract these motion elements. The analysis yields a generative motion model that can synthesize new motions in the distinctive styles of these individuals. The effectiveness of applying the tensor decomposition approach to our purpose was confirmed by synthesizing novel walking motions for a person by using the extracted signature.

    DOI

  • “Network Transmission of Human Face Data and 3D Reconstruction Considering Image Data Reduction”

    Hajime Sato, Jun Ohya, Nobuyoshi Terashima

    Journal of the Institute of Image Electronics Engineers of Japan   34 ( 5 ) 539 - 547  2005  [Refereed]

     View Summary

    A method for enabling efficient facial data transmission over the communication network and reconstructing realistic 3D face images is proposed. A conventional method which combines the advantages of the model-based and image-based approaches to reconstruct 3D facial images with high efficiency and precision on a standalone environment is extensively used. In our approach, the face image recognition part and the 3D image reconstruction part, which were conventionally carried out on the same PC, are separated and implemented on two network-connected PCs. Also, by using JPEG compression, the size of image data transmitted over the network is reduced. Our method was implemented on network-connected PCs, and its effectiveness was confirmed. Subjective tests showed that image data could be reduced to around 6% of its original size. © 2005, The Institute of Image Electronics Engineers of Japan. All rights reserved.

    DOI

  • Editorial - Special section on multimodal interfaces and applications

    J Ohya, R Reilly, CS Li

    IEEE TRANSACTIONS ON MULTIMEDIA   6 ( 3 ) 397 - 397  2004.06

    DOI

  • 強化現実技術を応用した立体ディスプレイシステムの開発

    盛川浩志, 川口真実, 河合隆史, 大谷淳

    日本人間工学会誌   第40巻特別号   372 - 373  2004.06

    DOI

  • Human motion tracking using nonsynchronous multiple observations

    H Mori, A Utsumi, J Ohya, M Yachida, R Nakatsu

    ELECTRONICS AND COMMUNICATIONS IN JAPAN PART III-FUNDAMENTAL ELECTRONIC SCIENCE   87 ( 12 ) 84 - 93  2004  [Refereed]

     View Summary

    This paper proposes a method of tracking a human object by using nonsynchronous multiple-viewpoint images. The proposed method tracks human forms efficiently by using a Kalman filter to integrate observed information which is obtained nonsynchronously from multiple viewpoints. The experimental system is composed of multiple observation nodes, which operate nonsynchronously to process the multiviewpoint images, a tracking node, which tracks the human figure, and the discovering node, which finds the human figure. The image features are matched to the tracking model in the observation node based on the prediction of the observed value which is sent from the tracking node. The image features matched to the model are sent to the tracking node, and the tracking model is updated. The image features which are not matched are sent to the discovering node to find a new human figure. With the proposed approach, it is possible to construct a large-scale tracking system while reducing the deterioration of processing efficiency and the redundancy among observations that occur in synchronous systems. The effectiveness of the proposed method is demonstrated by an experiment using real images. (C) 2004 Wiley Periodicals, Inc.

    DOI

  • Visual-dimension interact system (VIS) - Exhibiting creative process for museum visitor experience

    A Onda, T Oku, PY Chiu, E Yu, M Yokoi, Choh, I, J Ohya

    2004 INTERNATIONAL CONFERENCE ON CYBERWORLDS, PROCEEDINGS     278 - 285  2004  [Refereed]

     View Summary

    In this paper we describe a mixed reality supported interactive viewing enhancement museum display system. With a transparent interactive interface, the museum visitor is able to see, manipulate, and interact with the physical exhibit and its virtual information, which are overlapped on one other. Furthermore, this system provides the possibility for the visitor to experience the creation process in an environment as close as possible to the real process. This has the function of assisting the viewer in understanding the exhibit and most importantly, gaining a so-to-speak hands-on experience of the creation process itself leading to a deeper understanding of it.

  • Softness perceptive texture method

    Y Kobayashi, M Suzuki, J Ohya

    HUMAN VISION AND ELECTRONIC IMAGING IX   5292   322 - 329  2004  [Refereed]

     View Summary

    We have been studying about people's visual impression and image features for texture images in order to clarify the human subjective interpretation mechanism for images[1]. In corresponding image features of human impressions for the images, we found that the impressions for material were bottle-necked. We have studied a new analysis method which gives the impression for material from texture images. Especially, we mainly focused on the properties of visual targets which people can feel tactile sense. In this paper, we propose a new texture analysis method which is based on frequency analysis with 3D texture which is designed for photorealistic rendering. We found that our new method can estimate not only the surface roughness but also the surface softness.

  • Development of a stereoscopic 3D display system to observe restored heritage

    H Morikawa, M Kawaguchi, T Kawai, J Ohya

    STEREOSCOPIC DISPLAYS AND VIRTUAL REALITY SYSTEMS XI   5291   415 - 422  2004  [Refereed]

     View Summary

    The authors have developed a binocular-type display system that allows digital archives of cultural assets to be viewed in their actual environment. The system is designed for installation in locations where such cultural assets were originally present. The viewer sees buildings and other heritage items as they existed historically by looking through the binoculars. Images of the cultural assets are reproduced by stereoscopic 3D CG in cyberspace, and the images are superimposed on actual images in real-time. This system consists of stereoscopic CCD cameras that capture a stereo view of the landscape and LCDs for presentation to the viewer. Virtual cameras. used to render CG images from digital archives. move in synchrony with the actual cameras, so the relative position of the CG images and the landscape on which they are superimposed is always fixed. The system has manual controls for digital zoom. Furthermore, the transparency of the CG images can be altered by the viewer. As a case study for the effectiveness of this system, the authors chose the Heijyoukyou ruins in Nara, Japan. The authors evaluate the sense of immersion, stereoscopic effect, and usability of the system.

    DOI

  • Real-time stereo by using dynamic programming

    Sven Forstmann, Yutaka Kanou, Jun Ohya, Sven Thuering, Alfred Schmitt

    IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops   2004- ( January )  2004  [Refereed]

     View Summary

    A Dynamic Programming (DP) based algorithm that achieves real-time, high quality stereo-matching is presented. A special coarse to fine approach as well as the MMX based assembler implementation mainly contribute to the reached computation speed. A novel vertical smoothing approach by re-integrating paths inside the DP matching allows the avoidance of noisy horizontal strokes, so that high quality stereo-matching is achieved. The current implementation capable of running at about 30 FPS on an 2.2GHz PC, which is sufficient for the utilization in real-time applications.

    DOI

  • Exploiting the cognitive synergy between different media modalities in multimodal information retrieval

    ZF Zhang, RF Zhang, J Ohya

    2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3   CD-ROM proceedings (4 pages)   2227 - 2230  2004  [Refereed]

     View Summary

    This is a position paper reporting an on-going collaboration project between SUNY Binghamton, USA, and Waseda University, Japan, on multimodal information retrieval through exploiting the cognitive synergy across the different modalities of the information to facilitate an effective retrieval. Specifically we focus on image retrieval in the applications where imagery data appear along with collateral text. It is noted that these applications are ubiquitous. We have proposed the Synergistic Indexing Scheme (SIS) to explicitly exploit the synergy between the information of imagery and text modalities. Since the synergy we have exploited between the information of imagery and text modalities is subjective and depends on specific cognitive context, we call this type of synergy as cognitive synergy. We have reported part of the empirical evaluation and are in the process to fully implement the SIS prototype for an extensive evaluation.

  • Cognitive bridge between haptic impressions and texture images for subjective image retrieval

    Y Kobayashi, J Ohya, ZF Zhang

    2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3   CD-ROM proceedings (4 pages)   2239 - 2242  2004  [Refereed]

     View Summary

    As a step towards subjective image retrieval, this paper reports an on-going collaboration project between Waseda University and SUNY Binghamton, on relating texture images to haptic impressions. To grasp the surface height variations, texture images are taken under different illuminations and viewing conditions. Our method applies a new frequency analysis method to the texture images. We evaluate the performances of our feature and other typical conventional features by checking whether texture images are correctly classified into "soft" or "hard" by the SVM (support vector machine) method, where the training data for the SVM are collected by subjective tests. Experimental results show that our texture feature can classify "soft" or "hard" better than the other features.

  • Computer vision based analysis of the botanical tree's dynamical behaviors for the reproduction in virtual space

    LC Lu, J Ohya

    2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3   CD-ROM proceedings (4 pages)   839 - 842  2004  [Refereed]

     View Summary

    This paper deals with a method that analyzes a botanical tree's behaviors in real space by a computer vision approach so as to reproduce the analyzed behaviors in virtual space. Instead of applying unstable local tracking to the tree in a video sequence, we estimate the direction and strength of the wind that shakes the tree by a learning based method that classifies the input video sequence into one of the stored winds with different directions and strengths. In the learning phase, sample video sequences are used for constructing the Eigenspace and Fisherspace, which is obtained from Fisher discriminant analysis. In the classification phase, the input video sequence is compared with each of the stored sample sequences so that the direction and strength of the wind are estimated. An interpolation method improves the estimation accuracy. Experimental results demonstrate the effectiveness of the proposed method.

  • Realistic 3D facial animation using parameter-based deformation and texture remapping

    H Sato, J Ohya, N Terashima

    SIXTH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, PROCEEDINGS     735 - 742  2004  [Refereed]

     View Summary

    A method for constructing 3D human facial images which maintains the chararcteristics of video input image is proposed. First, a personalized 3D polygonal facial model is constructed from two orthogonal facial images. Video input frames are modified, and the vertices composing the 3D polygons are altered according to the contraction rule of facial muscles. Moreover, the modified input frame is texture mapped directly onto to the forementioned model to improve the quality of the reconstructed image. This method was implemented onto an ordinary PC, and an experiment using five persons' face data showed the effectiveness of the method both quantitatively and subjectively.

    DOI

  • Estimating complicated and overlapped human body postures by wearing a multiple-colored suit using color information processing

    DW Kang, Y Onuma, J Ohya

    SIXTH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, PROCEEDINGS     687 - 692  2004  [Refereed]

     View Summary

    This paper suggested an image processing method using a multiple-colored suit for estimating complicated postures of a human body. One of the major problems on the posture estimation when using the silhouette image analysis is the overlap of the body parts' silhouettes. In order to solve this problem, the human wears the multiple-colored suit so that each body part has a different color from the other body parts. Experimental results show the effectiveness of the proposed method. It has confirmed that the complicated overlapped postures can be analyzed through the contour of the human's silhouette and the color information from the body segments that are taken using three directional cameras which are located at the front, left, and right side of the subject.

    DOI

  • Real-time Method for Animating Elastic Objects' Behaviors Including Collisions

    The 13th International Conference on Artificial Reality and Telexistence     270 - 277  2003.12  [Refereed]

  • The Study of Communicability of Verbal and Non-Verbal expressions in Haiku and Haiga

    Proceedings of The 4th Japan-Korea International Symposium on Kansei Engineering     313 - 314  2003.10  [Refereed]

  • Computer vision based recognition of interactions between human body and object

    Masumi Kobana, Jun Ohya

    IFIP Advances in Information and Communication Technology   112   173 - 180  2003  [Refereed]

     View Summary

    This paper proposes a computer vision based method that recognizes interactions between human body and object. In two successive frames in a video sequence, our function based contour analysis method checks whether the silhouettes of the human body and object overlap. If they do not overlap, our method judges that the human body and object does not interact. If the two silhouettes overlap, our method checks whether the motion vectors obtained from the areas for the human body and object coincide. If they coincide, our method judges that the human body and object are interacting. Some experimental results show the effectiveness of the proposed method. © 2003 by Springer Science+Business Media New York.

    DOI

  • Estimating postures of a human wearing a multiple-colored suit based on color information processing

    DW Kang, J Ohya

    2003 INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL I, PROCEEDINGS   Ⅰ   261 - 264  2003  [Refereed]

     View Summary

    This paper proposes a non-contact type method for estimating human body postures. One of the major problems on the posture estimation using the silhouette image analysis is the overlapping of the body parts' silhouettes. In order to solve this problem, this paper proposes a method for estimating the posture of a human wearing a multiple-colored suit based on color information processing. By analyzing the contour of the human's silhouette, the method judges whether feature points are occluded by another body parts. If the occlusions occurs, color region segmentation is performed in order to know which region is frontal. The feature point in the frontal region is located in the skeleton of the region. Experimental results show the effectiveness of the proposed method.

    DOI

  • Efficient, realistic method for animating dynamic behaviors of 3D botanical trees

    H Kanda, J Ohya

    2003 INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL II, PROCEEDINGS   Ⅱ   89 - 92  2003  [Refereed]

     View Summary

    This paper proposes anew efficient method that can animate botanical trees in 3D realistically. In this paper, a 3D botanical tree model consists of a set of branch segments, to which leaf models are attached. To reduce the amount of computation, instead of calculating the motions of all the branch segments, only the representative segment in each branch is numerically analyzed. The numerical analysis is constrained to a 2D plane so that 3D numerical analysis need not be performed. Concerning the leaf model, a set of four leaves is systematically attached to each branch segment. Experimental results clarify the conditions for real-time, realistic animations of dynamic behaviors of trees.

    DOI

  • Construction of realistic three dimensional facial images using video images

    H Sato, J Ohya, N Terashima

    SEVENTH INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND ITS APPLICATIONS, VOL 2, PROCEEDINGS     227 - 230  2003  [Refereed]

     View Summary

    A method for constructing 3D human facial images which maintains the chararcteristics of video input image is proposed. First, a personalized 3D polygonal facial model is constructed from two orthogonal facial images. Video input frames are modified, and the movements of facial feature points are tracked and measured based on the the contraction rule of facial muscles. The vertices of the 3D polygons are altered accordingly, and the modified input frame is texture mapped directly onto to the forementioned model. This method was implemented using an ordinary CCD camera and computer, and an experiment using one person's data showed that realistic facial images could be reconstructed with high speed and precision.

    DOI

  • Analyzing Video Sequences of Multiple Humans - Tracking, Posture Estimation and Behavior Recognition.

    Jun Ohya, Akira Utsumi, Junji Yamato

      3   i-xxii, 1-138  2002  [Refereed]

    DOI

  • Face posture estimation using eigen analysis on an IBR (image based rendered) database

    Kuntal Sengupta, Philip Lee, Jun Ohya

    Pattern Recognition   35 ( 1 ) 103 - 117  2002.01  [Refereed]

     View Summary

    In this paper, we present a novel representation of the human face for estimating the orientation of the human head in a two dimensional intensity image. The method combines the use of the much familiar eigenvalue based dissimilarity measure with image based rendering. There are two main components of the algorith described here: The offline hierarchical image database generation and organization, and the online pose estimation stage. The synthetic images of the subject's face are automatically generated offline, for a large set of pose parameter values, using an affine coordinate based image reprojection technique. The resulting database is formally called as the IBR (or image based rendered) database. This is followed by the hierarchical organization of the database, which is driven by the eigenvalue based dissimilarity measure between any two synthetic image pair. This hierarchically organized database is a detailed, yet structured, representation of the subject's face. During the pose estimation of a subject in an image, the eigenvalue based measure is invoked again to search the synthetic (IBR) image closest to the real image. This approach provides a relatively easy first step to narrow down the search space for complex feature detection and tracking algorithms in potential applications like virtual reality and video-teleconferencing applications. © 2001 Pattern Recognition Society. Published by Elsevier Science Ltd. All rights reserved.

    DOI

  • Construction of facial expressions using a muscle-based feature model

    YC Liu, H Sato, N Terashima, J Ohya

    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL I AND II, PROCEEDINGS   CD-Rom Proceedings 4 pages   469 - 472  2002  [Refereed]

     View Summary

    An efficient method for constructing facial images for use in telecommunication applications is proposed. This method uses a simple 3D feature model, which consists of polygons, which describe the shape of the face, and elastic linear springs, which simulate the natural movements of facial muscles. This method requires only two orthogonal facial images, and could easily be implemented on a relatively low-spec PC. Experimental results showed good results that various facial expressions could be synthesized and displayed from arbitrary directions.

  • Analysis of human behaviors by computer vision based approaches

    J Ohya

    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL I AND II, PROCEEDINGS   CD-Rom Proceedings 4 pages   913 - 916  2002  [Refereed]

     View Summary

    This paper describes the author's recent activities related to computer vision based methods for analyzing human behaviors: more specifically, posture estimation and recognizing interactions between a human body and object. For estimating postures in 3D from multiple camera images, the authors developed a heuristic based method and non-heuristic method. The heuristic based method heuristically analyzes the contour of a human silhouette so that significant points of a human body can be located in each image. The non-heuristic method utilizes a function for analyzing contours without using heuristic rules. Recognizing the interactions exploits the function based contour analysis and motion vector based analysis so that the system can judge whether the human body interacts with the object.

  • Spatial filtering using the active-space indexing method

    SK Semwal, J Ohya

    GRAPHICAL MODELS   63 ( 3 ) 135 - 150  2001.05  [Refereed]

     View Summary

    Traditionally most camera-based position estimation systems use only a few points to calibrate cameras. In this paper, we investigate a novel and alternate approach for 3D position estimation by using a larger number of points arranged in a 3D grid. We present an implementation of the active-space indexing mechanism which uses three cameras. Given the corresponding points in camera images, a precise estimation of the position can be obtained. The active-space indexing method can be also used as a spatial filter to eliminate the large number of possible corresponding pairs from consideration. This capability, unique only to the active-space indexing method, provides a tractable algorithm to the otherwise intractable situation. (C) 2001 Academic Press.

    DOI

  • User-Guided Composition Effects For Art-Based Rendering

    2001 ACM Symposium on Interactive 3D Graphics, ACM SIGGRAPH     99 - 102  2001.03  [Refereed]

  • 非同期多視点画像による人物追跡システムの構築

    電子情報通信学会論文誌D-II   Vol. J84-D-II ( No.1 ) 102 - 110  2001.01  [Refereed]

  • Computer vision based analysis of non-verbal information in HCI

    Jun Ohya

    Proceedings - IEEE International Conference on Multimedia and Expo     869 - 872  2001  [Refereed]

     View Summary

    This paper overviews our research activities on computer vision based non-verbal information analysis that can be applied to virtual communication environments and human computer interactions. In virtual communication environments, a user's facial expressions and body motions are estimated by computer vision approaches, and the estimated non-verbal information is reproduced in the user's avatar. For human computer interfaces, hand gestures are recognized as pre-defined commands by analyzing multiple camera images that observe the hand. In addition, facial expressions and body gestures are recognized from a time-sequential images by HMM (Hidden Markov Models).

    DOI

  • Remarks on a real-time, noncontact, nonwear, 3D human body posture estimation method

    Kazuhiko Takahashi, Tatsumi Sakaguchi, Jun Ohya

    Systems and Computers in Japan   31 ( 14 ) 1 - 10  2000.12  [Refereed]

     View Summary

    This paper proposes a new real-time method of estimating human postures in three dimensions from trinocular images. The proposed method extracts feature points of the human body by analyzing contours of human silhouettes. The feature points are extracted by using the subtraction images when self-occlusions occur in the silhouette images. Dynamic compensation is carried out with a Kalman filter so that all feature points are tracked. The 3D coordinates of the feature points are reconstructed by considering the geometrical relationship between the three cameras. Experimental results confirm both the feasibility and the effectiveness of the proposed method.

    DOI

  • Computer Vision Based Analysis of Non-verbal Information

    Jun Ohya

    Sixth International Conference on Control, Automation, Robotics and VIsion   CD-ROM Proceedings 6 pages  2000.12  [Refereed]

  • Virtual Environments for Human-to-Human Communication

    Jun Ohya

    IEEE International Conference on Multimedia and Expo   Tutorial 4B Notes  2000.07  [Refereed]

  • Bunshin: A Believable Avatar Surrogate for Both Scripted and On-the-Fly Pen-Based Control in a Presentation Environment

    Sketches & Applications, SIGGRAPH2000 Conference Abstracts and Applications     187  2000.07  [Refereed]

  • Remarks on Real-Time Human Body Posture Estimation Using Neural Networks

    K. Takahashi, T. Uemura and Jun Ohya

    6th International Conference on Engineering Applications of Neural Networks     229 - 236  2000.07  [Refereed]

  • Epipolar constraint from 2D affine lines, and its application in face image rendering

    K Sengupta, J Ohya

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E83D ( 7 ) 1567 - 1573  2000.07  [Refereed]

     View Summary

    This paper has two parts. In the first part of the paper, we note the property that under the para perspective camera projection model of a camera, the set of 2D images produced by a 3D point can be optimally represented by two lines in the affine space (alpha - beta space). The slope of these two lines are same, and we observe that this constraint is exactly the same as the epipolar line constraint. Using this constraint, the equation of the epipolar line can be derived. In the second part of the paper, we use the "same slope" property of the lines in the alpha - beta space to derive the affine structure of the human face. The input to the algorithm is not limited to an image sequence of a human head under rigid motion. It can be snapshots of the human face taken by the same or different cameras, over different periods of time. Since the depth variation of the human face is not very large, we use the para perspective camera projection model. Using this property, we reformulate the (human) face structure reconstruction problem in terms of the much familiar multiple baseline stereo matching problem. Apart from the face modeling aspect. we also show how we use the results for reprojecting human faces in identification tasks.

  • 実時間非接触非装着三次元人物姿勢推定方法に関する一考察

    高橋和彦, 坂口竜己, 大谷 淳

    電子情報通信学会論文誌 D-Ⅱ   J83-D-Ⅱ ( 5 ) 1305 - 1314  2000.05  [Refereed]

  • 人物像の実時間認識・生成

    システム制御情報学会誌 システム/制御/情報   Vol.44 ( No.3 ) 135 - 140  2000.03  [Refereed]

  • Two-step approach for real-time eye tracking with a new filtering technique

    S Kawato, J Ohya

    SMC 2000 CONFERENCE PROCEEDINGS: 2000 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOL 1-5     1366 - 1371  2000  [Refereed]

     View Summary

    Head and face detection and eye tracking in realtime are the first steps for head gesture recognition and/or face expression recognition for a human-computer interaction interface. We propose a two-step approach for eye tracking in video streams.
    First, we detect or track a point between the eyes. For this task, we apply a special filter rye proposed before. Once we detect the point between the eyes, it is fairly easy to locate the eyes, which are the two small darkest parts on each side of this point. Because defecting the point between the eves is easier and more stable than directly detecting the eyes, the system can robustly track the eyes
    We implement the system on a SGI O2,workstation. The video image size is 320x240 pixels. The system processes images at seven frames per second in the detection mode, and 13 frames per second in the tracking mode without arty special hardware.

  • Locating human face in a complex background including non-face skin colors

    JS Tang, S Kawato, J Ohya, R Nakatsu

    CISST'2000: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGING SCIENCE, SYSTEMS, AND TECHNOLOGY, VOLS I AND II     117 - 122  2000  [Refereed]

     View Summary

    This paper proposes a skin color information based face location algorithm that can locate the human face in a complex background including non-face skin color regions, such as hands. Our algorithm is divided into three steps. The first step is to use color components of the input image to segment the skin colors from the complex background. The second step is to detect a curve, which is called the division curve, between the hair region and face region. The last step is to use the division curve to estimate the face location. Our algorithm can be applied to some gesture analysis and understanding systems that combine the gestures of the head and hands.

  • Automatic skin-color distribution extraction for face detection and tracking

    S Kawato, J Ohya

    2000 5TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS, VOLS I-III     1415 - 1418  2000  [Refereed]

     View Summary

    For face detection and tracking in a video sequence, many systems are now applying a technique to extract skin-color areas effaces. In these systems, the skin-color model should be built beforehand. However skin-color changes when the lighting condition varies when an input video camera is changed, when the white-balance of the camera is readjusted, etc.
    In this paper we describe a method to build a skin-color model for operation while a face detection system is running. When it has no skin-color model, it uses the differences between adjacent frames to extract moving objects. Once the system recognizes that the moving object is a face, it analyses the color histograms of the moving urea and extract skin-color distribution
    We applied this method to our face detection and tracking system. Experiments showed good results.

  • Adaptive human motion tracking using non-synchronous multiple viewpoint observations

    A Utsumi, H Yang, J Ohya

    15TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 4, PROCEEDINGS     607 - 610  2000  [Refereed]

     View Summary

    In this paper, we propose and adaptive human tracking system with non-synchronous multiple observations. Our system consists of three types of processes, discovering node for detecting newly appeared person, tracking node for tracking each target person, and observation node for processing one viewpoint (camera) images. We have multiple observation nodes and each node works fully independently. The tracking node integrates observed information based on reliability evaluation. Both observation conditions (number of cameras), relative distance between a human and cameras, extent of occlusion, etc.) and human motion states (walking, standing, sitting) are considered in the evaluation. Matching between tracking models and observed image features are performed in each observation node based on the position, size and color similarities of each 2D image. Due to the non-synchronous property, this system is highly scalable for increasing the detection area and number of observing nodes. Experimental results for some indoor scenes are also described.

    DOI

  • Human body postures from trinocular camera images

    Shoichiro Iwasawa, Jun Ohya, Kazuhiko Takahashi, Tatsumi Sakaguchi, Kazuyuki Ebihara, Shigeo Morishima

    Proceedings - 4th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2000     326 - 331  2000  [Refereed]

     View Summary

    This paper proposes a new real-time method for estimating human postures in 3D from trinocular images. In this method, an upper body orientation detection and a heuristic contour analysis are performed on the human silhouettes extracted from the trinocular images so that representative points such as the top of the head can be located. The major joint positions are estimated based on a genetic algorithm-based learning procedure. 3D coordinates of the representative points and joints are then obtained from the two views by evaluating the appropriateness of the three views. The proposed method implemented on a personal computer runs in real-time. Experimental results show high estimation accuracies and the effectiveness of the view selection process. © 2000 IEEE.

    DOI

  • Real-time detection of nodding and head-shaking by directly detecting and tracking the "between-eyes"

    Shinjiro Kawato, Jun Ohya

    Proceedings - 4th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2000     40 - 45  2000  [Refereed]

     View Summary

    Among head gestures, nodding and head-shaking are very common and used often. Thus the detection of such gestures is basic to a visual understanding of human responses. However it is difficult to detect them in real-time, because nodding and head-shaking are fairly small and fast head movements. We propose an approach for detecting nodding and head-shaking in real time from a single color video stream by directly detecting and tracking a point between the eyes, or what we call the "between-eyes". Along a circle of a certain radius centered at the "between-eyes", the pixel value has two cycles of bright parts (forehead and nose bridge) and dark parts (eyes and brows). The output of the proposed circle-frequency filter has a local maximum at these characteristic points. To distinguish the true "between-eyes" from similar characteristic points in other face parts, we do a confirmation with eye detection. Once the "between-eyes" is detected, a small area around it is copied as a template and the system enters the tracking mode.combining with the circle-frequency filtering and the template, the tracking is done not by searching around but by selecting candidates using the template
    the template is then updated. Due to this special tracking algorithm, the system can track the "between-eyes" stably and accurately. It runs at 13 frames/s rate without special hardware. By analyzing the movement of the point, we can detect nodding and head-shaking. Some experimental results are shown. © 2000 IEEE.

    DOI

  • Multiple-Camera-Based Human Tracking using Non-synchronous Observations

    Fourth Asian Conference on Computer Vision     1034 - 1039  2000.01  [Refereed]

  • Neural-network-based real-time human body posture estimation

    K Takahashi, T Uemura, J Ohya

    NEURAL NETWORKS FOR SIGNAL PROCESSING X, VOLS 1 AND 2, PROCEEDINGS     477 - 486  2000  [Refereed]

     View Summary

    This paper proposes a real-time human body posture estimation method using ANNs. The network is composed of three ANNs and a decision logic unit. The ANNs' input is the result of a function analysis on a human silhouette's contour extracted from camera images and the ANNs' output indicates the feature points' positions on the contour. The decision logic unit synthesizes each of the ANNs' output vectors and then the 2D coordinates of the human body's feature points are calculated. The proposed method is implemented on a personal computer and runs in real-time (17-20 frames/sec). Experimental results confirm both the feasibility and the effectiveness of the proposed method for estimating human body postures.

  • Remarks on a real-time 3D human body posture estimation method using trinocular images

    K Takahashi, T Sakaguchi, J Ohya

    15TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 4, PROCEEDINGS   Vol. 4   693 - 697  2000  [Refereed]

     View Summary

    This paper proposes a new real-time method of estimating human postures in 3D from trinocular images. The proposed method extracts feature points Of the human body by applying a type of function analysis to contours of human silhouettes. To overcome self-occlusion problems, dynamic compensation is carried out using the Kalman filter and all feature points are tracked. The 3D coordinates of the feature points are reconstructed by considering the geometrical relationship between the three cameras. Experimental results confirm both the feasibility and the effectiveness of the proposed method, and an application example of the 3D human body posture estimation to a motion recognition system is presented.

  • Virtual Me: A virtual communication method that enables simultaneous multiple existence as an avatar and/or agents

    J Ohya, R Nakatsu, S Kawato, T Sakaguchi

    2000 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, PROCEEDINGS VOLS I-III     1657 - 1660  2000  [Refereed]

     View Summary

    This paper proposes a new communication method, called "Virtual Me", that enables a user to simultaneously participate in multiple Events such as meetings or lectures as a virtual-metamorphosed avatar controlled by the user and/or as a computer-controlled autonomous agents. By interpreting non-verbal data generated by the other participants at each event, the system determines when to switch between using virtual metamorphosis or employing agents. We demonstrate the effectiveness of this method in a prototype of the Virtual Me system. It is our hope that this technology may be in the interest not only of time efficiency but also of enriched human communication.

  • Modeling and Animation of Botanical Tree for Interactive Virtual Environments

    Symposium on Virtual Reality Software and Technology 1999    1999.12  [Refereed]

  • Indexing method for three-dimensional position estimation

    Fermin, I, S Semwal, J Ohya

    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS   E82D ( 12 ) 1597 - 1604  1999.12  [Refereed]

     View Summary

    Indexing techniques usually are used in model-based object recognition and ray tracing algorithms. In this paper we present a new method for estimating the three-dimensional position of a subject (resp. object) in a circumscribed space based on an indexing method. We construct two and three-dimensional indices of a space, which are used to estimate the three-dimensional position by an interpolation technique. There are two processes in estimating the three-dimensional position of a subject (resp. object): preprocessing and three-dimensional position estimation. We have implemented this idea using stereo camera, and tested by using two different sizes of a grid pattern. Promising results for preprocessing and 3D position estimation are presented. Moreover, we show that this approach can also be extended for multiple cameras.

  • A Face Recognition System Based on Wavelet Transform and Neural Network

    International conference on wavelet analysis and its applications    1999.11  [Refereed]

  • A New Size-Limited Filter Bank for Finite-signal Subband Processing

    International conference on wavelet analysis and its applications    1999.11  [Refereed]

  • Face Detection from a Complex Background

    International Workshop on Very Low Bitrate Video Coding    1999.10  [Refereed]

  • Real-time computer vision based technologies for capturing human motions

    14th International Symposium on Computer and Information Sciences    1999.10  [Refereed]

  • Virtual Metamorphosis Systems

    1999 IEEE International Conference on Systems, Man, and Cybernetics    1999.10  [Refereed]

  • Comparison of Neural-Network-Based Pattern Classification Methods with Application to Human Motion Recognition

    Fifth International Conference on Engineering Applications of Neural Networks (EANN '99)    1999.09  [Refereed]

  • 実写映像に基づいた3次元樹木モデルの生成(共著)

    坂口竜己, 大谷 淳, 中津良平

    電子情報通信学会論文誌   J82-D-Ⅱ ( 9 ) 1469 - 1477  1999.09  [Refereed]

  • Towards the realization of non-verbal communication

    HCI'99 (Human-Computer Interaction)   Vol.2  1999.08  [Refereed]

  • 多数カメラを用いた両手手振りの検出

    情報処理学会論文誌   Vol.40, No.8  1999.08  [Refereed]

  • Multiple-Hand-Gesture-Tracking using Multiple Cameras

    IEEE Computer Society Conference on Computer Vision and Pattern Recogntion (CVPR99)    1999.06  [Refereed]

  • Real-time 3-D facial image reconstruction for virtual space teleconferencing

    K Ebihara, N Suzuki, J Ohya, F Kishino

    ELECTRONICS AND COMMUNICATIONS IN JAPAN PART III-FUNDAMENTAL ELECTRONIC SCIENCE   82 ( 5 ) 80 - 90  1999.05  [Refereed]

     View Summary

    In this paper, a new method was proposed for implementing 3-D facial image models designed to allow faithful reconstruction of facial images in virtual space teleconferencing by using 3-D measurement while detecting various facial expressions. In the method proposed, first many dots are painted on the face. Then, for a variety of facial expressions (eight expressions in this study) selected with regard to the action of the major facial muscles, 3-D displacement vectors for the normal (neutral) face are measured at each dot and recorded in the facial image plane as reference vectors. When the facial image is reconstructed, 2-D displacement vectors are detected by tracking markers that were previously placed on the face and are represented as the sum of two enclosing reference vectors. Based on these data, the vertices of a 3-D wireframe face model (WFM) are moved appropriately, and the facial expression is reconstructed. (C) 1999 Scripta Technica.

  • Virtual metamorphosis

    J Ohya, J Kurumisawa, R Nakatsu, K Ebihara, S Iwasawa, D Harwood, T Horprasert

    IEEE MULTIMEDIA   6 ( 2 ) 29 - 39  1999.04  [Refereed]

     View Summary

    The virtual metamorphosis system lets people change their forms into any other form in a virtual scene. To realize these changes, a computer Vision system estimates facial expressions and body postures and reproduces them in a computer graphics avatar in real time. We introduce three systems in order of their development: the Virtual Kabuki system, Networked Theater, and "Shall We Dance?".

    DOI

  • 多数カメラを用いた手形状認識法とその仮想空間インタフェースへの応用

    内海 章, 大谷 淳, 中津良平

    情報処理学会論文誌   40 ( 2 ) 585 - 593  1999.02  [Refereed]

  • Virtual reality technologies for multimedia communications

    J Ohya, T Miyasato, R Nakatsu

    MIXED REALITY     285 - 300  1999  [Refereed]

  • Extracting facial motion parameters by tracking feature points

    T Otsuka, J Ohya

    ADVANCED MULTIMEDIA CONTENT PROCESSING   1554   433 - 444  1999  [Refereed]

     View Summary

    A method for extracting facial motion parameters is proposed. The method consists of three steps. First, the feature points of the face, selected automatically in the first frame, are tracked in successive frames. Then, the feature points are connected with Delaunay triangulation so that the motion of each point relative to the surrounding points can be computed. Finally, muscle motions are estimated based on motions of the feature points placed near each muscle. The experiments showed that the proposed method can extract facial motion parameters accurately. In addition, the facial motion parameters are used to render a facial animation sequence.

  • Real-time estimation of human body postures using Kalman filter

    K Takahashi, T Sakaguchi, J Ohya

    RO-MAN'99: 8TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN INTERACTION     189 - 194  1999  [Refereed]

     View Summary

    This paper presents a hybrid estimation method of human body postures from CCD camera images. In the hybrid estimation method, the feature points of the human body (top of the head, tips of the hands, and feet, and elbow joints) are obtained from the results of heuristic contour analyses of human silhouettes or those of a time subtraction image depending on the reliability of the silhouette information. A dynamic compensation is then carried out by tracking all feature points using the AR model in order to obtain their optimal position and to overcome self-occlusion problems. The AR model's parameters are estimated through on-line processing by the Kalman filter. The proposed method is implemented on a personal computer and the process runs in real-time. Experimental results show high estimation accuracy and the feasibility of the proposed method.

  • Real-time, 3D estimation of human body postures from trinocular images

    Shoichiro Iwasawa, Jun Ohya, Kazuhiko Takahashi, Tatsumi Sakaguchi, Sinjiro Kawato, Kazuyuki Ebihara, Sigeo Morishima

    Proceedings - IEEE International Workshop on Modelling People, MPeople 1999     3 - 10  1999  [Refereed]

     View Summary

    This paper proposes a new real-time method for estimating human postures in 3D from trinocular images. In this method, an upper body orientation detection and a heuristic contour analysis are performed on the human silhouettes extracted from the trinocular images so that representative points such as the top of the head can be located. The major joint positions are estimated based on a genetic algorithm based learning procedure. 3D coordinates of the representative points and joints are then obtained from the two views by evaluating the appropriateness of the three views. The proposed method implemented on a personal computer runs in real-time (30 frames/second). Experimental results show high estimation accuracies and the effectiveness of the view selection process.

    DOI

  • A real-time surveillance and evaluation system of the audience's reaction for meeting sessions

    JS Tang, J Ohya

    INTELLIGENT ROBOTS AND COMPUTER VISION XVIII: ALGORITHMS, TECHNIQUES, AND ACTIVE VISION   3837   45 - 57  1999  [Refereed]

     View Summary

    In this paper, we describe a visual surveillance system for evaluating the audience's reaction in meeting sessions. The system, which works in real-time, can recognize and evaluate the reaction of the audience. It is mainly composed of three subsystems. The first subsystem is a face detection and head motion segmentation system which is used to detect the face from complex background and segment the head motion into different units, with each unit including different information about the audience's reaction. The second subsystem is a gesture and pose recognition system which can recognize the gestures and the poses of human head. The third subsystem is an evaluation system which is used to evaluate the reaction of the audience by using the recognition results in the second system. Our system has been tested for determining the concentration of the audience in meeting sessions and the experimental results are good.

  • Artistic anatomy based, real-time reproduction of facial expressions in 3D face models

    J Ohya, K Ebihara, J Kurumisawa

    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA COMPUTING AND SYSTEMS, PROCEEDINGS VOL 2   Vol.2   684 - 689  1999  [Refereed]

     View Summary

    This paper proposes a new real-time method for reproducing facial expressions in 3D face models realistically based on anatomy for artists. To reproduce facial expressions in a face model, the detected expressions need to be converted to the data for deforming the face model. In the proposed method, an artist who has learned anatomy for artists creates arbitrary facial expressions in the 3D face model by mixing the reference expressions chosen by the artist so that the synthesized expressions realistically represent the respective expressions displayed by real persons. The parameters obtained by this manual operations are used to construct the equations that convert the expression features obtained by the detection module to the displacement vectors of the vertices of the face model. During human communications through face models, the equations are used to reproduce the detected expressions in real-time. The effectiveness and robustness of the proposed method were demonstrated by experimental results and demonstration systems.

  • Novel scene generation, merging and stitching views using the 2D affine space

    K Sengupta, J Ohya

    SIGNAL PROCESSING-IMAGE COMMUNICATION   14 ( 1-2 ) 39 - 53  1998.11  [Refereed]

     View Summary

    In this paper we present a unified theoretical framework for novel scene synthesis, merging real and virtual worlds, and view stitching. To start with, we have a set of real images from weakly calibrated cameras, for which we compute the dense point match correspondences. For applications like novel view synthesis, one may first solve the 3D scene reconstruction problem, followed by a view rendering process. However, errors in 3D scene reconstruction usually gets reflected in the quality of the new scene generated, so we seek. a more direct method. In this paper, we use the knowledge of dense point matches and their affine coordinate values to estimate the corresponding affine coordinate values in the new scene. Our technique of reprojection is extended for other applications like merging real and synthetic worlds, and view stitching. (C) 1998 Elsevier Science B.V. All rights reserved.

    DOI

  • Face/gesture analysis/synthesis telchnologies for humanto-human communications through virtual environments

    Jun Ohya

    Proceedings of the 6th ACM International Conference on Multimedia: Face/Gesture Recognition and their Applications, MULTIMEDIA 1998     12 - 19  1998.09  [Refereed]

     View Summary

    This paper describes the vutual metamorphosis system, whtch was originally proposed by ATR Media Integratton &amp
    Communicattons Research Laboratories, Kyoto, Japan. In the vntual metamorphosis system, people can change their forms mto any other form in a vutual scene. To realize. this changes, facial expressions and body postures are estimated by computer vision and reproduced m a computer graphics avatar m real-ume. We developed computer vision based technologies for automatic face trackmg, estimatmg deformations of factal components,and estimatmg body postures in real-time. We also developed real-time facial expresston reproduction method based on "Anatomy for artists". Usmg these technologies, prototyping systems for virtual metamorpho-SIS were constructed. This paper introduces two systems in order of their development: the Virtual Kabuki system and the "Shall We Dance" system.

    DOI

  • 画素値分布の推定による時系列画像からの動物体領域抽出

    内海 章, 大谷 淳, 中津良平

    電子情報通信学会論文誌 D-II   J81-D-II ( 8 ) 1768 - 1775  1998.08  [Refereed]

  • 美術解剖学に基づいた仮想変身システムのための実時間人物表情再現(共著)

    海老原一之, 楜沢 順, 大谷 淳, 中津良平, 岸野文郎

    電子情報通信学会論文誌   J81-D-Ⅱ ( 5 ) 841 - 849  1998.05  [Refereed]

  • 複雑な輪郭の自動抽出・追跡法

    畑幸一, 大谷 淳, 岸野文郎, 中津良平

    電子情報通信学会論文誌 D-II   .J81-D-II ( 4 ) 706 - 715  1998.04  [Refereed]

  • A new robust real-time method for extracting human silhouettes from color images

    Masanori Yamada, Kazuyuki Ebihara, Jun Ohya

    Proceedings - 3rd IEEE International Conference on Automatic Face and Gesture Recognition, FG 1998     528 - 533  1998  [Refereed]

     View Summary

    A new method for extracting human silhouettes from color images in real-time regardless of changes in lighting conditions and backgrounds is proposed. In the method, RGB values of each pixel in color images acquired by a CCD camera are converted to YIQ values. Before extracting human silhouettes, the mean YIQ values of each pixel and the threshold values for the Y, I, Q components are obtained in advance from the image sequence acquired by the camera that observed the background. Based on the mean values and the threshold values, each pixel is judged whether it belongs to the human silhouette or not. The proposed method was used with the virtual metamorphosis system, in which anyone can change his or her form to any other form, and the robustness and real-time-ness were confirmed. © 1998 IEEE.

    DOI

  • Spotting segments displaying facial expression from image sequences using HMM

    Takahiro Otsuka, Jun Ohya

    Proceedings - 3rd IEEE International Conference on Automatic Face and Gesture Recognition, FG 1998     442 - 447  1998  [Refereed]

     View Summary

    A method that can be used for spotting segments that display facial expression is proposed. The motion of the face is modeled by HMM in such a way that each state corresponds to the conditions of facial muscles, e.g., relaxed, contracting, apex and relaxing. The probability assigned to each state is updated iteratively as the feature vector is obtained from image processing. A spotted segment is placed into a certain category when the probability of that category exceeds a threshold value. Experiments show that the segments for the six basic expressions can be spotted accurately in near real time. © 1998 IEEE.

    DOI

  • A new camera projection model and its application in reprojection

    K Sengupta, J Ohya

    FOURTEENTH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOLS 1 AND 2     59 - 63  1998  [Refereed]

     View Summary

    In this paper we present a new camera projection model, which is intermediate between the affine camera mode and the pin hole projection model. It is modeled as a perspective projection of 3D points into an arbitrary plane, followed by an affine transform of these projected points. We observe that the reprojection of a point into a navel image can be achieved uniquely provided that we have located a set of Jive reference points over four images (of which three are input images, and the fourth is the novel image). Also, the reprojection, theory; does not assume that the input images are captured from cameras with identical internal calibration parameters. Thus, we apply our technique two different domain: (a) generation of novel images from a stereo pair (b) generation of virtual walkthrough sequence with a monocular image sequence as input.

  • Geometric-imprints: A significant points extraction method for the scan & track virtual environment

    SK Semwal, J Ohya

    AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS     480 - 485  1998  [Refereed]

     View Summary

    We have implemented a new algorithm, called the geometric-imprints method for extracting key-features of a given contour. Series of points on the curve are used to find a geometric-imprint set of the contour. The algorithm is based upon the assumption that an articulated human and other figures have mostly cylindrical body-parts. The algorithm recursively partitions the contour based upon the cylindrical nature of human-silhouettes, and a starting point. Similar topological contours are expected to produce similar geometric-imprints. Thus the correspondence problem is somewhat simplified. Results of our implementation are presented.

  • Human face structure estimation from multiple images using the 2D affine space

    K Sengupta, J Ohya

    AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS     106 - 111  1998  [Refereed]

     View Summary

    In this paper we present an algorithm to estimate the human face structure. The input to the algorithm is not limited to an image sequence of a human head under rigid motion. It can be snapshots of the human face taken by the same or different cameras, over different periods of time. Since the depth variation of the human face is not very large, we use the affine camera projection model. Under this assumption, it can be shown that the set of 2D images produced by a 3D point feature of a rigid object can be optimally represented by two lines in the affine space. Using this property, we reformulate the (human) face structure reconstruction problem in terms of the much familiar multiple baseline stereo matching problem [8]. Apart from the face modeling aspect, we also show how we use the results for reprojecting human faces in identification tasks.

  • Generating virtual environments for human communications - Virtual metamorphosis system and novel view generation

    J Ohya, K Sengupta

    COMPUTER VISION FOR VIRTUAL REALITY BASED HUMAN COMMUNICATIONS - 1998 IEEE AND ATR WORKSHOP PROCEEDINGS     43 - 50  1998  [Refereed]

     View Summary

    At ATR Media Integration & Communications Research Laboratories, our research group is aiming at realizing virtual communication environments in which the users at remote sites can feel that they are virtually co-located. This paper presents the ongoing research in two subareas chosen to address this problem: (a) Generating human images in virtual scenes (b) Act image based approach for generation of (background) scenes. In the virtual metamorphosis system, 3D models of characters are created in advance. For metamorphosis, the facial expressions and body posture of a person are detected in real-time fr om the face images and monocular thermal images of the person, respectively. The detected expressions and body movements are reproduced by deforming the 3D models. Demonstrations using SD models of a Kabuki actor and a dinosaur show good performance. for novel view generation, we use as affine coordinate based re-projection scheme, which is also extended to merging real and synthetic objects.

  • Real-time human posture estimation using monocular thermal images

    S Iwasawa, K Ebihara, J Ohya, S Morishima

    AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS     492 - 497  1998  [Refereed]

     View Summary

    This paper introduces a new real-lime method to estimate the posture of a human from thermal images acquired by an infrared camera regardless of the background and lighting conditions. Distance transformation is performed for the human body area extracted from the thresholded thermal image for the calculation of the center of gravity. After the orientation of the zipper half of the body is obtained by calculating the moment of inertia, significant points such as the top of the head, the tips oft he hands and foot are heuristically located. In addition, the elbow and knee positions are estimated from the detected (significant) points using a genetic algorithm based learning procedure.
    The experimental results demonstrate the robustness of the proposed algorithm and real-time (faster than 20 frames per second) performance.

  • Multiple-view-based tracking of multiple humans.

    Akira Utsumi, Hiroki Mori, Jun Ohya, Masahiko Yachida

    Fourteenth International Conference on Pattern Recognition, ICPR 1998, Brisbane, Australia, 16-20 August, 1998     597 - 601  1998  [Refereed]

    DOI

  • Direct Manipulation Interface Using Multiple Cameras for Hand Gesture Recognition.

    Akira Utsumi, Jun Ohya

    IEEE International Conference on Multimedia Computing and Systems, ICMCS 1998, Austin, Texas, USA, June 28 - July 1, 1998     264 - 267  1998  [Refereed]

    DOI

  • Image segmentation for human tracking using sequential-image-based hierarchical adaptation

    Akira Utsumi, Jun Ohya

    Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition     911 - 916  1998  [Refereed]

     View Summary

    We propose a novel method of extracting a moving object region from each frame in a series of images regardless of complex, changing background using statistical knowledge about the target. In vision systems for 'real worlds' like a human motion tracker, a priori knowledge about the target and environment is often limited (e.g., only the approximate size of the target is known) and is insufficient for extracting the target motion directly. In our approach, information about both target object and environment is extracted with a small amount of given knowledge about the target object. Pixel value (color, intensity, etc.) distributions for both the target object and background region are adaptively estimated from the input image sequence based on the knowledge. Then, the probability of each pixel being associated with the target object is calculated. The target motion can be extracted from the calculated stochastic image. We confirmed the stability of this approach through experiments.

    DOI

  • Multiple Camera Based Human Motion Estimation.

    Akira Utsumi, Hiroki Mori, Jun Ohya, Masahiko Yachida

    Computer Vision - ACCV'98, Third Asian Conference on Computer Vision, Hong Kong, China, January 8-10, 1998, Proceedings, Volume II     655 - 662  1998  [Refereed]

    DOI

  • 距離変換処理を用いた多視点画像による手姿勢推定法(共著)

    内海 章, 宮里 勉, 岸野文郎, 大谷 淳, 中津良平

    映像情報メディア学会誌   51 ( 12 ) 2116 - 2125  1997.12  [Refereed]

    DOI

  • 遺伝的アルゴリズムを用いた顔画像からの表情と顔の姿勢の推定

    大谷 淳

    電子情報通信学会論文誌   J80-D-Ⅱ ( 8 ) 2152 - 2160  1997.08  [Refereed]

  • 連続出力確率密度分布を用いたHMMによる動画像からの複数人物の表情認識

    大塚尚宏, 大谷 淳, 中津良平

    電子情報通信学会論文誌   J80-D-Ⅱ ( 8 ) 2129 - 2137  1997.08  [Refereed]

  • An affine coordinate based algorithm for reprojecting the human face for identification tasks

    K Sengupta, J Ohya

    INTERNATIONAL CONFERENCE ON IMAGE PROCESSING - PROCEEDINGS, VOL III     340 - 343  1997  [Refereed]

     View Summary

    In this paper we present an algorithm to generate new views of a human face, starting with at least two other views of the face. In a typical face recognition system, the task of comparison becomes easier if the faces have similar orientation with respect to the camera. The affine coordinate based reprojection algorithm presented in the paper enables us to do that. Dense point matches between the two input faces of the same individual are computed using an affine coordinate based reprojection framework. This is followed by the reprojection of one of these to faces to the target face once the user has matched four feature points across two input face images and the target face image.

  • Novel scene generation, merging and stitching views using the 2D affine space

    K Sengupta, J Ohya

    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA COMPUTING AND SYSTEMS '97, PROCEEDINGS     602 - 603  1997  [Refereed]

     View Summary

    In this paper we present an algorithm to generate new views of a scene, starting from stereo images. Errors in 3D scene reconstruction usually get reflected in the quality of the new scene generated, so we seek a direct method for reprojection. In this paper, we use the knowledge of dense point matches and their affine coordinate values to estimate the corresponding affine coordinate values in the new scene.

  • Hand Image Segmentation Using Sequential-Image-Based Hierarchical Adaptation.

    Akira Utsumi, Jun Ohya

    Proceedings 1997 International Conference on Image Processing, ICIP '97, Santa Barbara, California, USA, October 26-29, 1997     208 - 211  1997  [Refereed]

    DOI

  • Real-time estimation of human body postures from monocular thermal images

    Shoichiro Iwasawa, Kazuyuki Ebihara, Jun Ohya, Ryohei Nakatsu, Shigeo Morishima

    Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers   51 ( 8 ) 1270 - 1277  1997  [Refereed]

     View Summary

    This paper proposes a new real-time method for estimating human body postures from thermal images acquired by an infrared camera, regardless of the background and lighting conditions. Distance transformation is performed for the human body area extracted from the thresholded thermal image, in order to calculate the center of gravity. After the orientation of the upper half of the body is obtained by calculating the moment of inertia, significant points such as the top of the head and the ends of the hands and feet are heuristically located. In addition, the elbow and knee positions are estimated from the detected (significant) points, using a genetic-algorithm-based learning procedure. The experimental results demonstrate the robustness of the proposed algorithm and real-time performance (faster than 20 frames per second).

    DOI

  • Model-based estimation of human posture parameters from multiple camera images using genetic algorithms

    Jun Ohya, Fumio Kishino

    Kyokai Joho Imeji Zasshi/Journal of the Institute of Image Information and Television Engineers   51 ( 12 ) 2107 - 2115  1997  [Refereed]

     View Summary

    A new method for estimating human posture from multiple images using a genetic algorithm is proposed. In the proposed algorithm, the posture parameters to be estimated are assigned to the genes of an individual in the population. For each individual, its fitness evaluates to what extent the human multiple images synthesized by deforming a 3D human model according to the values of the genes are registered to the real human multiple images. Genetic operations such as natural selection, crossover and mutation are performed so that individuals in the next generation are generated. After a certain number of repetitions of these processes, the estimated parameter values are obtained from the individual with the best fitness. Experiments using multiple synthesized images show promising results for estimating 17 joint angle values for each degree of freedom of the joints and also the three translational and three rotational degrees of freedom.

    DOI

  • 臨場感通信会議のための実時間表情検出

    海老原一之, 大谷 淳, 岸野文郎

    テレビジョン学会誌   50 ( 10 ) 1497 - 1506  1996.10  [Refereed]

    DOI

  • 形状の面対称性に基づく能動的観測による3次元物体形状の復元および色彩情報の再現

    吉田和憲, 大谷 淳, 岸野文郎, 田中弘美

    テレビジョン学会誌   50 ( 10 ) 1536 - 1542  1996.10  [Refereed]

    DOI

  • 臨場感通信会議のための実時間3次元表情再現

    海老原一之, 鈴木紀子, 大谷 淳, 岸野文郎

    電子情報通信学会論文誌A   J79-A ( 2 ) 527 - 536  1996.02  [Refereed]

  • Detecting facial expressions from face images using a genetic algorithm

    Jun Ohya, Fumio Kishino

    Proceedings - International Conference on Pattern Recognition   3   649 - 653  1996  [Refereed]

     View Summary

    A new method to detect deformations of facial parts from a face image regardless of changes in the position and orientation of a face using the genetic algorithm is proposed. Facial expression parameters that are used to deform and position a 3D face model are assigned to the genes of an individual in a population. The face model is deformed and positioned according to the gene values of each individual and is observed by a virtual camera, and a face image is synthesized. The fitness which evaluates to what extent the real and synthesized face images are similar to each other is calculated. After this process is repeated for sufficient generations, the parameter estimation is obtained from the genes of the individual with the best fitness. Experimental results demonstrate the effectiveness of the method. © 1996 IEEE.

    DOI

  • Automatic extraction and tracking of contours

    Koichi Hata, Jun Ohya, Fumio Kishino, Ryohei Nakatsu

    Proceedings - International Conference on Pattern Recognition   1   441 - 445  1996  [Refereed]

     View Summary

    This paper considers the problem of extracting and tracking complex contours without user interaction. We assume that a complex contour consists of contour segments whose spatial coordinates and intensity gradient vary smoothly in the direction normal to themselves. In our algorithm, digital curves that could correspond to contour segments are extracted by connecting edge pixels using a B-spline based contour segment model. The extracted curves trace the contour segments at the next frame by using the active contour model technique. Experimental results show even occluded contours can be tracked automatically. © 1996 IEEE.

    DOI

  • Real-time facial expression detection based on frequency domain transform

    K Ebihara, J Ohya, F Kishino

    VISUAL COMMUNICATIONS AND IMAGE PROCESSING '96   2727   916 - 926  1996  [Refereed]

  • Virtual Kabuki Theater: Towards the realization of human metamorphosis systems

    J Ohya, K Ebihara, J Kurumisawa, R Nakatsu

    RO-MAN '96 - 5TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS     416 - 421  1996  [Refereed]

     View Summary

    This paper describes the Virtual Kabuki Theater the authors have recently developed. In the Virtual Kabuki Theater, people at different locations can be Kabuki actors in a Kabuki scene. In our system, Kabuki actors' SD models are created in advance. Facial expressions of a person are detected an real-tame in the face images from the small camera fixed to the helmet worn by the person. Body movements of the person are estimated in real-time from the thermal images acquired by the infrared camera that observes the person. The detected expressions and body movements are reproduced in. the Kabuki actor's model. Our implementation shows good performance. The Virtual Kabuki Theater is a first step towards human metamorphosis systems, in which anyone can change (metamorphose) his/her form into any other characters.

  • Pose estimation of human body part using multiple cameras

    K Sengupta, J Ohya

    RO-MAN '96 - 5TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS     146 - 151  1996  [Refereed]

     View Summary

    In this paper, we present a method of obtaining the approximate transformation parameter values as a starting point in estimating the pose of rigid 3D free form objects using multiple 2D images. We back project the edge silhouettes in the images, and obtain the approximate volume in the 3D space containing the object. Next, for a point selected in the volume, we hypothesize a set of points within the 3D CAD model of the object it can possibly correspond to, using the spatial extent function introduced in this paper. This is repeated for three arbitrarily chosen point in the volume. The hypothesized (match point) lists of these three points are next used to derive the pose parameter by enforcing the conditions of rigidity. Our initial experiments demonstrate the potential of this idea, and the pose parameters estimated using this method can be refined using the standard methods available in the literature.

  • Realistic and real-time reproduction of human body movements for virtual space teleconferencing

    Keiichi Nariyama, Karansher Singh, Jun Ohya, Fumio Kishino

    Terebijon Gakkaishi (Journal of the Institute of Television Engineers of Japan)   50 ( 9 ) 1352 - 1362  1996  [Refereed]

     View Summary

    This paper describes realistic, real-time 3-D reproduction of human body movements based on detection data from a small number of sensors attached to a participant for virtual space teleconferencing. A 3-D human model is created by a skeletal link model. WFM (Wire Frame Model), body segments for FFD (Free Form Deformations), and IFP (Implicit Function Primitives) for body parts. In this model WFM provides fast rendering ability and FFD and IFP give natural deformation representation. A reconstruction algorithm that can deal with human movements in teleconference situations is proposed
    sensed data are used to reconstruct whole body movements by utilizing motion constraints. Experiments, in which four magnetic sensors for detecting the six pose parameters are attached to the upper half of a human body, show the effectiveness of the proposed 3-D human model and reconstruction method, in which human movements are reproduced at a speed of 8 frames/sec.

    DOI

  • 隠れマルコフモデルによる顔動画像からの表情認識

    坂口竜己, 大谷 淳, 岸野文郎

    テレビジョン学会誌   49 ( 8 ) 1060 - 1067  1995.08  [Refereed]

    DOI

  • VIRTUAL SPACE TELECONFERENCING - REAL-TIME REPRODUCTION OF 3D HUMAN IMAGES

    J OHYA, Y KITAMURA, F KISHINO, N TERASHIMA, H TAKEMURA, H ISHII

    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION   6 ( 1 ) 1 - 25  1995.03  [Refereed]

     View Summary

    Real-time reproduction of 3D human images is realized by an experimental system recently built as a prototype for virtual space teleconferencing, in which participants at different sites can feel as if they are colocated and can work cooperatively, At each sending and receiving site of the teleconferencing system, a 3D model of each participant is constructed from a wire frame model mapped by color texture and is rendered on a 3D display, In the current experimental system, real-time detection of facial features at the sending site is achieved by visually tracking tape marks pasted to the participant's face. Movements of the head, body, hands, and fingers are detected in real time using magnetic sensors and data gloves. At the receiving site, the detected motion parameters are used to move nodes in the wire frame model to reproduce the movements of the participants at each sending site. Realistic facial expressions are reproduced through the-use of simple motion rules applied to the tape mark tracking information, Through experimental evaluation, the optimal number of nodes for best quality has been obtained. Reproduction of facial expressions and synthesis of arm movements are examined, The reproduction speed using the optimal human model is approximately 6 frames/s. Examples of cooperative work by participants using the experimental system illustrate the effectiveness of virtual space teleconferencing. (C) 1995 Academic Press, Inc.

    DOI

  • Real Time Facial Expression Detection Based On Frequency Domain Transform

    EBIHARA Kazuyuki, OHYA Jun, KISHINO Fumio

    Proceedings of The ITE Winter Annual Convention   1995   82 - 82  1995

    DOI CiNii

  • A study of real time facial expression detection for virtual space teleconferencing

    K Ebihara, J Ohya, F Kishino

    RO-MAN'95 TOKYO: 4TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS     247 - 252  1995  [Refereed]

  • Realistic 3D facial animation in virtual space teleconferencing

    L Moubaraki, J Ohya, F Kishino

    RO-MAN'95 TOKYO: 4TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN COMMUNICATION, PROCEEDINGS     253 - 258  1995  [Refereed]

  • Active 3D modeling by recursive viewpoint selection based on symmetry

    K Yoshida, HT Tanaka, J Ohya, F Kishino

    INTELLIGENT ROBOTS AND COMPUTER VISION XIV: ALGORITHMS, TECHNIQUES, ACTIVE VISION, AND MATERIALS HANDLING   2588   326 - 336  1995  [Refereed]

  • RECOGNIZING CHARACTERS IN SCENE IMAGES

    J OHYA, A SHIO, S AKAMATSU

    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE   16 ( 2 ) 214 - 220  1994.02  [Refereed]

     View Summary

    An effective algorithm for character recognition in scene images is studied. Scene images are segmented into regions by an image segmentation method based on adaptive thresholding. Character candidate regions are detected by observing gray-level differences between adjacent regions. To ensure extraction of multisegment characters as well as single-segment characters, character pattern candidates are obtained by associating the detected regions according to their positions and gray levels. A character recognition process selects patterns with high similarities by calculating the similarities between character pattern candidates and the standard patterns in a dictionary and then comparing the similarities to the thresholds. A relaxational approach to determine character patterns updates the similarities by evaluating the interactions between categories of patterns, and finally character patterns and their recognition results are obtained. Highly promising experimental results have been obtained using the method on 100 images involving characters of different sizes and formats under uncontrolled lighting.

    DOI

  • DENSE, TIME-VARYING RANGE DATA-ACQUISITION FROM STEREO PAIRS OF THERMAL AND INTENSITY IMAGES

    J OHYA, F KISHINO

    1994 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, PROCEEDINGS     860 - 865  1994  [Refereed]

  • HUMAN POSTURE ESTIMATION FROM MULTIPLE IMAGES USING GENETIC ALGORITHM

    J OHYA, F KISHINO

    PROCEEDINGS OF THE 12TH IAPR INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION - CONFERENCE A: COMPUTER VISION & IMAGE PROCESSING     750 - 753  1994  [Refereed]

  • 隠れマルコフモデルを用いた動画像からの人物の行動認識(共著)

    大和淳司, 大谷 淳, 石井健一郎

    電子情報通信学会論文誌   J76-D-Ⅱ ( 12 ) 2556 - 2563  1993.12  [Refereed]

    CiNii

  • 3次元形状の軸表現:Smoothed Local Generalized Cones

    佐藤嘉伸, 大谷 淳, 石井健一郎

    電子情報通信学会論文誌D-II   J76-D-II ( 8 ) 1612 - 1623  1993.08  [Refereed]

  • 距離画像からの3次元形状の階層的パーツ構造の復元

    佐藤嘉伸, 大谷 淳, 石井健一郎

    電子情報通信学会論文誌D-II   J76-D-II ( 8 ) 1624 - 1636  1993.08  [Refereed]

  • A NEW METHOD FOR ACQUIRING TIME-SEQUENTIAL RANGE IMAGES BY INTEGRATING STEREO PAIRS OF THERMAL AND INTENSITY IMAGES

    J OHYA, F KISHINO

    1993 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION : PROCEEDINGS     730 - 731  1993  [Refereed]

  • VIRTUAL SPACE TELECONFERENCING SYSTEM - REAL-TIME DETECTION AND REPRODUCTION OF 3-D HUMAN IMAGES

    F KISHINO, J OHYA, H TAKEMURA, N TERASHIMA

    HUMAN-COMPUTER INTERACTION, VOL 2   19   669 - 674  1993  [Refereed]

  • REAL-TIME REPRODUCTION OF 3D HUMAN IMAGES IN VIRTUAL SPACE TELECONFERENCING

    J OHYA, Y KITAMURA, H TAKEMURA, F KISHINO, N TERASHIMA

    IEEE VIRTUAL REALITY ANNUAL INTERNATIONAL SYMPOSIUM     408 - 414  1993  [Refereed]

  • TIME-VARYING HOMOTOPY AND THE ANIMATION OF FACIAL EXPRESSIONS FOR 3D VIRTUAL SPACE TELECONFERENCING

    S KAJIWARA, HT TANAKA, Y KITAMURA, J OHYA, F KISHINO

    VISUAL COMMUNICATIONS AND IMAGE PROCESSING 93, PTS 1-3   2094   37 - 48  1993  [Refereed]

  • 多次元心理空間を用いる類似画像検索法

    田邊勝義, 大谷 淳, 石井健一郎

    電子情報通信学会論文誌D-II   J75-D-II ( 11 ) 1856 - 1865  1992.11  [Refereed]

    CiNii

  • FACIAL IMAGE SYNTHESIS BY HIERARCHICAL WIRE FRAME MODEL

    Y KITAMURA, Y NAGASHIMA, J OHYA, F KISHINO

    VISUAL COMMUNICATIONS AND IMAGE PROCESSING 92, PTS 1-3   1818   1358 - 1365  1992  [Refereed]

  • 情景画像からの文字パターンの抽出と認識

    大谷 淳, 塩 昭夫

    電子情報通信学会論文誌D   J71-D ( 6 ) 1037 - 1047  1988.06  [Refereed]

  • リングパターン法を用いるロボット用物体認識センサ

    大谷 淳

    電子通信学会論文誌D   J69-D ( 3 ) 269 - 278  1986.03  [Refereed]

  • GRAY-SCALE PRINTING ON PLAIN PAPER USING THERMAL INK-TRANSFER IMAGING

    J OHYA, Y TOKUNAGA

    JOURNAL OF IMAGING TECHNOLOGY   10 ( 2 ) 57 - 63  1984  [Refereed]

  • COLOR THERMAL INK-TRANSFER IMAGING

    J OHYA, Y TOKUNAGA

    IEEE TRANSACTIONS ON ELECTRON DEVICES   31 ( 9 ) 1258 - 1263  1984  [Refereed]

    DOI

  • NEW GRAY-SCALE PRINTING METHOD USING A THERMAL PRINTER

    Y TOKUNAGA, K KUBOTA, J OHYA

    IEEE TRANSACTIONS ON ELECTRON DEVICES   30 ( 8 ) 898 - 904  1983  [Refereed]

    DOI

▼display all

Books and Other Publications

  • Analyzing video sequences of multiple humans: Tracking, posture estimation and behavior recognition

    Kluwer Academic Publishers  2002.03

Misc

  • 森林環境モニタリングロボットにおけるSLAMと深層学習を用いた領域推定に基づく経路計画法の検討

    森本純矢, 林拓哉, 藤本裕之, 石井裕之, 大谷淳, 大和淳司, 高西淳夫

    画像電子学会年次大会予稿集(CD-ROM)   46th   ROMBUNNO.S1‐3  2018.06

    J-GLOBAL

  • 環境モニタリングロボットの視覚情報処理を用いた森林における起点への帰還法の検討

    藤本裕之, 石井裕之, 大谷淳, 大和淳司, 高西淳夫

    画像電子学会年次大会予稿集(CD-ROM)   45th   ROMBUNNO.S‐2  2017.06

    J-GLOBAL

  • H-012 Study of Ultrasound Image Segmentation Based on SuperVoxel for Extracting Fetal Tissues

    Wang Yuanchen, Tatematsu Naotomo, Ohya Jun, Xu Rong, Ishikawa Makiko, Iwata Hiroyasu

      14 ( 3 ) 149 - 150  2015.08

    CiNii

  • H-008 Study of SLAM Based Confirmation of Results of Mechanical Operations by a Clean-Up Robot

    GAO Si, TATEMATSU Naotomo, OHYA Jun

      14 ( 3 ) 135 - 136  2015.08

    CiNii

  • H-030 A Method for Tracking Guitar Neck and Fingertips: Necking Tracking Robust against Occlusions Based on Geometry Analysis and Fingertips Tracking Based on Temporal Probability Map

    Zhao Wang, Ohya Jun

      14 ( 3 ) 195 - 196  2015.08

    CiNii

  • I-015 Study of an Automatic System that Generates Portraits from Facial Images Using Neural Networks : Comparative Study Based on Subjective Tests

    GUO Shuai, ENOMOTO Makoto, KAWAMURA Harumi, OHYA Jun

      14 ( 3 ) 231 - 232  2015.08

    CiNii

  • I-009 Study of Tracking the Entire Tree from a Video Sequence Based on Tracking a Small Number of Feature Points

    Shen Yuanchao, Ohya Jun

      14 ( 3 ) 217 - 218  2015.08

    CiNii

  • D-16-12 Fundamental Study of Early Recognition of Surgeon's Action during Suture Surgery from Video Sequences

    Li Ye, Ohya Jun, Chiba Toshio, Xu Rong, Yamashita Hiromasa

    Proceedings of the IEICE General Conference   2015 ( 2 ) 152 - 152  2015.02

    CiNii

  • D-11-38 Fundamental Study of Tracking Fingers in RGBD Dynamic Image Sequences towards Automatic Teaching System for Piano Practice

    Qiu Dong, Ohya Jun

    Proceedings of the IEICE General Conference   2015 ( 2 ) 38 - 38  2015.02

    CiNii

  • D-12-53 Study of Early Event Recognition by Utilizing Skeletons Based on Relative Positions of Joints of Human Bodies in a Lie Group and Max-Margin Early Event Detectors

    Tatematsu Naotomo, Ohya Jun, Davis Larry S.

    Proceedings of the IEICE General Conference   2015 ( 2 ) 105 - 105  2015.02

    CiNii

  • D-11-12 Study of a Vision Based Method for Checking the Position of Each Finger of Guitar Players : Towards the Actualization of an Autonomous Guitar Chord Teaching System

    WANG Zhao, OHYA Jun

    Proceedings of the IEICE General Conference   2015 ( 2 ) 12 - 12  2015.02

    CiNii

  • 抽象絵画における鑑賞者の視線行動を利用した音楽生成システムの研究-鑑賞者の発話内容の重複と発想の広がりとの関連に関する検討-

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    電子情報通信学会大会講演論文集(CD-ROM)   2015  2015

    J-GLOBAL

  • 抽象絵画における鑑賞者の注視情報を利用した音楽生成システム-抽象絵画の特徴が鑑賞者の抽象的な発想の変化に及ぼす影響-

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    情報科学技術フォーラム講演論文集   14th  2015

    J-GLOBAL

  • Study of Recognizing Customers' Behaviors in Shops by Dynamic Image Processing and Bayesian Network

    YAN Jing, HORI Takayuki, OHYA Jun

    Technical report of IEICE. PRMU   114 ( 197 ) 101 - 106  2014.09

     View Summary

    This paper explores the validity of embedding the image processing based extraction of 10 image features that could describe customers' shopping behaviors to a Bayesian Network based classification method that can recognize the four categories: combinations of with/without hesitation and bought/not bought. By constructing a system that consists of the three modules for tracking customers, extracting the image features and classifying the customers' behaviors, recognition tests using 48 video data were conducted. As a result of the experiments, a recognition accuracy of higher than 89% is achieved, which is promising.

    CiNii

  • Study of Recognizing Customers' Behaviors in Shops by Dynamic Image Processing and Bayesian Network

    Jing Yan, Takayuki Hori, Jun Ohya

    IPSJ SIG Notes. CVIM   2014 ( 17 ) 1 - 6  2014.08

     View Summary

    This paper explores the validity of embedding the image processing based extraction of 10 image features that could describe customers' shopping behaviors to a Bayesian Network based classification method that can recognize the four categories: combinations of with/without hesitation and bought/not bought. By constructing a system that consists of the three modules for tracking customers, extracting the image features and classifying the customers' behaviors, recognition tests using 48 video data were conducted. As a result of the experiments, a recognition accuracy of higher than 89% is achieved, which is promising.

    CiNii

  • I-036 Study of Subjective Discrimination in an Automatic System for Generating Line Drawing Based Portraits from Facial Images

    Wu Yuzhen, Enomoto Makoto, Ohya Jun

      13 ( 3 ) 247 - 248  2014.08

    CiNii

  • H-027 Study of Image Features for a Bayesian Network Based Method for Recognizing Customers' Behaviors in Shops

    Yan Jing, Hori Takayuki, Ohya Jun

      13 ( 3 ) 121 - 122  2014.08

    CiNii

  • I-035 Fundamental Study of Virtual Make-up System : Study of Retrieving Made-up Faces Based on Similarities of Face Shapes

    Zhang Lin, Ohya Jun

      13 ( 3 ) 245 - 246  2014.08

    CiNii

  • Study of Portrait Similarities in the Automatic System that Generates Portraits from Facial Images

    Wu Yuzhen, Enomoto Makoto, Ohya Jun

    IEICE technical report. Image engineering   114 ( 172 ) 1 - 6  2014.08

     View Summary

    This paper studies an automatic method that generates portraits by utilizing a neural network trained by facial features obtained by applying Active Shape Model to original face images and by portraits drawn by an illustrator. The authors' conventional research works had remaining issues such as low subjective similarities of generated portraits, validity of the nose, and absence of subjective evaluation of generated portraits. To increase the subjective similarity, this paper proposes a new portrait depicting tool. We obtain promising feeling that subjective similarity can be improved by the new depicting tool. In addition, we conduct experiments on subjective discrimination of portraits. It turns out that the new tool gives higher subjective discrimination rate than our conventional tool, and that the case of the new tool and presence of the nose gives the highest discrimination rate.

    CiNii

  • Study of Clustering Feature points of Moving Objects From Dynamic RGB-D Images : Proposing a Method that Combines Supervoxel and Multi Label Graph-Cut

    Tatematsu Naotomo, Ohya Jun, Davis Larry S.

    Technical report of IEICE. PRMU   113 ( 493 ) 127 - 132  2014.03

     View Summary

    This paper proposes a method that can discriminate feature points of each moving object from the still background by utilizing supervoxel and Multi-label Graph-Cut. The ordinal clustering method based on multi label graph-cut has a problem that can not classify small movement object. To solve this problem, our method utilize initial cluster based on supervoxel and construct multi-labeled-graph by the unit of group that belongs same supervoxel. This improvement enable accurate clustering feature points even if the amount of movement is small. Experiments using multiple moving objects and real stereo sequences demonstrate the effectiveness of our proposed method.

    CiNii

  • Study of Recognizing Hand Actions from Video Sequences during Suture Surgeries Based on Temporally-Sectioned SIFT and Sliding Window Based Neural Networks

    Li Ye, Ohya Jun, Chiba Toshio, Xu Rong, Yamashita Hiromasa

    Technical report of IEICE. PRMU   113 ( 493 ) 151 - 156  2014.03

     View Summary

    Towards the realization of a robotic nurse that can support surgeries autonomously by recognizing surgical situations only using video informations, this paper proposes an improved method by using sectioned-SIFT and sliding window based neural network that can recognize surgeon's hand actions: suture and tying. Hand area is detected by using color information and then the video sequence is partitioned into sections. Sectioned-SIFT descriptors are computed in each section and built a word vocabulary. Histogram feature of the action is spliced by using word's frequency in each section. Finally, sliding window and neural network is used to recognize the significant actions: suture and tying. The proposed method has achieved the 100% recognition rate for manually extracted actions and 90% recognition rate for whole surgery video sequences.

    CiNii

  • Fundamental Study of Recognizing Customers' Behaviors in Shops Based on Facial Expression Analysis : Acquiring Facial Expression Features by ASM and How to Build Bayesian Networks' Models

    JIA Luwei, OHYA Jun, HORI Takayuki

    Technical report of IEICE. PRMU   113 ( 493 ) 157 - 162  2014.03

     View Summary

    Towards the realization of a method that can recognize the psychological state that causes the purchasing behaviors of consumers, this paper proposes a method for recognizing customers' behaviors based on analyses of the customers' facial expressions. First, in each frame of the video sequence, facial feature points are extracted by ASM (Active Shape Model), and features are obtained from feature points in the eye-brows and mouth. Then, each feature is assigned to a node of Bayesian Network. On recognition, by inputting a video sequence, the Bayesian Network outputs the behavior recognition result. The validity of the proposed method is presented experimentally.

    CiNii

  • Study of Tongue Detection and Tongue Shape Classification Based on Active Contour Model and Texture Features

    TIAN Ming, LUO Dan, OHYA Jun

    Technical report of IEICE. PRMU   113 ( 493 ) 133 - 138  2014.03

     View Summary

    In this paper, we aim to an automatic tongue diagnosis system in traditional Chinese medicine. First, we propose a new method to extract the tongue area automatically from digital photos using the active contour model based on the watershed segmentation. Second, we classify the following symptoms: the teeth-printed tongue, which could be diagnosed by the contour of tongue, the tongue coating, the fissured tongue, which cloud be diagnosed by the texture feature of tongue, as well as the swelling tongue or thin tongue, which could be diagnosed by the area of tongue. Finally, we confirm the effectiveness of the proposed method by experiments.

    CiNii

  • Study of the Optimal Number and Locations of Facial Feature Points in Generating Portraits Using Neural Networks

    FANG Guoyong, ENOMOTO Makoto, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   113 ( 470 ) 145 - 150  2014.03

     View Summary

    We study a system that can automatically generate portraits from face images using multi-layered neural networks, for which image features of the face detected by ASM are the inputs, the portrait drawn by an illustrator are teaching signals, and back-propagation algorithms are performed for learning. This system aims at outputting portraits that do not depend on deformations of the line drawing traced from the face image, and captures the drawing style of an individual artist. The number and locations of feature points detected by ASM could affect quality of generated portraits. This paper reports new results that was not clear in our laboratory conventional research. We study the optimal number and locations of the detected facial feature points in generating portraits.

    CiNii

  • D-12-76 Fundamental Study of Recognizing Customers' Entire Bodies' Behaviors in Shops Using Image Processing and Bayesian Network

    Yan Jing, Hori Takayuki, Ohya Jun

    Proceedings of the IEICE General Conference   2014 ( 2 ) 151 - 151  2014.03

    CiNii

  • D-16-5 Study of Recognizing Surgeon's Action during Suture Surgery by Using Sectioned-SIFT

    Li Ye, Ohya Jun, Chiba Toshio, Xu Rong, Yamashita Hiromasa

    Proceedings of the IEICE General Conference   2014 ( 2 ) 169 - 169  2014.03

    CiNii

  • D-12-68 Study of Lip Reading from Video Sequences by Combining Optical Flow Features and Shape Features with Discrete Cosine Transform Features

    Takahashi Shohei, Ohya Jun

    Proceedings of the IEICE General Conference   2014 ( 2 ) 143 - 143  2014.03

    CiNii

  • D-12-45 Study of Tongue Detection and Tongue Shape Classification for Tongue Diagnosis

    Tian Ming, Luo Dan, Ohya Jun

    Proceedings of the IEICE General Conference   2014 ( 2 ) 120 - 120  2014.03

    CiNii

  • D-12-65 Study of Clustering Moving Objects's Feature points by Supervoxel and Multi-label Graph-Cut

    Tatematsu Naotomo, Ohya Jun, Davis Larry S.

    Proceedings of the IEICE General Conference   2014 ( 2 ) 140 - 140  2014.03

    CiNii

  • D-12-75 Study of Recognizing Customers' Behaviors Based on Facial Expression Analysis and Bayesian Networks

    Jia Luwei, Ohya Jun, Hori Takayuki

    Proceedings of the IEICE General Conference   2014 ( 2 ) 150 - 150  2014.03

    CiNii

  • D-11-5 Study of an Automatic Line Drawing Based Portrait Generation System Based on Neural Network : Study of Generating Noses

    Wu Yuzhen, Enomoto Makoto, Ohya Jun

    Proceedings of the IEICE General Conference   2014 ( 2 ) 5 - 5  2014.03

    CiNii

  • D-11-6 Fundamental Study of Generating Portraits Using Neural Networks : Study of the Optimal Number of Facial Feature Points Based on Physical Errors between the Feature Points and Generated Portraits

    Fang Guoyong, Enomoto Makoto, Ohya Jun

    Proceedings of the IEICE General Conference   2014 ( 2 ) 6 - 6  2014.03

    CiNii

  • Automatic Lip-Reading by using Multiple Visual Features -Integration of the Shape, Optical Flaw and DCT features-

    Shohei Takahashi, Jun Ohya

    IPSJ SIG Notes. CVIM   2014 ( 7 ) 1 - 7  2014.02

     View Summary

    In the paper, we present a lip-reading method that can recognize speech by using only visual features. Lip-reading can work well in noisy places such as in the car or in the train. In addition people with hearing-impaired or difficulties in hearing can be benefited. First, the Active Shape Model (ASM) is applied to track and detect the face and lip in a video sequence. Second, three visual features, the shape, optical flow and Discreet cosine transformation of the lip are obtained from the lip area detected by ASM. The extracted features are ordered chronologically so that Support Vector Machine (SVM) is performed so as to learn and classify the spoken words. Hierarchical SVMs are used to recognize the words. Each visual feature is trained by the respective middle-layer SVM, and those outputs of SVM's are integrated by the final SVM. Experimental results show that the integration of these features improves the recognition accuracy.

    CiNii

  • 抽象絵画の鑑賞者の視線行動に基づく音楽生成システムの予備的検討-音楽の有無が抽象絵画鑑賞者の発想の多寡に与える影響-

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    電子情報通信学会大会講演論文集(CD-ROM)   2014  2014

    J-GLOBAL

  • Fundamental Study of a Music Generation System Based on Gaze Behaviors of Viewers of Abstract Paintings-Effects of Proposed System on the Amount of Ideas of Viewers-

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    電子情報通信学会技術研究報告   113 ( 501(HIP2013 78-105) )  2014

    J-GLOBAL

  • 抽象絵画における鑑賞者の視線行動を利用した音楽生成システムの研究-鑑賞者の発想を膨らませる音楽の自動生成システムに関する予備的検討-

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    情報科学技術フォーラム講演論文集   13th  2014

    J-GLOBAL

  • Study of Structure-from-motion Based 3D Reconstruction and Surface Mosaic

    Duan Mengyao, Xu Rong, Ohya Jun

      113 ( 197 ) 113 - 118  2013.09

     View Summary

    Structure-from-motion (SFM) is a significant issue of 3D reconstruction for many applications. However, some image collector like endoscopy only provides a narrow view angle and 2D structures of the internal organs. In order to expand the view angle and describe the details of 3D structures, we propose an approach to reconstruct 3D surface based on image stitching from a video sequence. There are four major steps of the proposed method: 3D reconstruction, 3D point cloud stitching, dense point cloud generation and Poisson surface reconstruction. Experimental results demonstrate that the proposed method achieves a satisfied accuracy and efficiency for 3D reconstruction, and promise a large view angle.

    CiNii

  • Study of Recognizing Spontaneous Facial Expressions of a Person who Watches Web News Based on ASM and Bayesian Network

    XU Chao, OHYA Jun

      113 ( 197 ) 7 - 12  2013.09

     View Summary

    To achieve a comment system that can automatically estimate emotions of a person who watches web news, this paper proposes a computer vision based method that can recognize facial expressions generated in response to Web news. During the learning session, first, ASM (Active Shape Model) is applied to the face so that some feature points are located. From the located feature points, some features such as distances between feature points are obtained. The features collected from sample data are used for nodes of Bayesian Networks. During facial expression estimation session, the features that are computed from the input facial video sequence by the same procedure as the learning session are input to the Bayesian Networks so that the recognition result is obtained as the facial expression category that maximizes the a-posteriori probability. Experimental results demonstrate the effectiveness of the proposed method.

    CiNii

  • Study of Recognizing Spontaneous Facial Expressions of a Person who Watches Web News Based on ASM and Bayesian Network

    XU Chao, OHYA Jun

    Technical report of IEICE. PRMU   113 ( 196 ) 7 - 12  2013.09

     View Summary

    To achieve a comment system that can automatically estimate emotions of a person who watches web news, this paper proposes a computer vision based method that can recognize facial expressions generated in response to Web news. During the learning session, first, ASM (Active Shape Model) is applied to the face so that some feature points are located. From the located feature points, some features such as distances between feature points are obtained. The features collected from sample data are used for nodes of Bayesian Networks. During facial expression estimation session, the features that are computed from the input facial video sequence by the same procedure as the learning session are input to the Bayesian Networks so that the recognition result is obtained as the facial expression category that maximizes the a-posteriori probability. Experimental results demonstrate the effectiveness of the proposed method.

    CiNii

  • Study of Structure-from-motion Based 3D Reconstruction and Surface Mosaic

    Duan Mengyao, Xu Rong, Ohya Jun

    Technical report of IEICE. PRMU   113 ( 196 ) 113 - 118  2013.09

     View Summary

    Structure-from-motion (SFM) is a significant issue of 3D reconstruction for many applications. However, some image collector like endoscopy only provides a narrow view angle and 2D structures of the internal organs. In order to expand the view angle and describe the details of 3D structures, we propose an approach to reconstruct 3D surface based on image stitching from a video sequence. There are four major steps of the proposed method: 3D reconstruction, 3D point cloud stitching, dense point cloud generation and Poisson surface reconstruction. Experimental results demonstrate that the proposed method achieves a satisfied accuracy and efficiency for 3D reconstruction, and promise a large view angle.

    CiNii

  • I-046 Fundamental Study of Generating Portraits Using Neural Networks : Study of Relationship between Face Features Detected by ASM and the Generated Portrait Quality

    Fang Guoyong, Enomoto Makoto, Ohya Jun

      12 ( 3 ) 347 - 348  2013.08

    CiNii

  • H-034 Fundamental Study of Recognizing Customers' Entire Bodies' Behaviors in Shops Using Image Processing and Bayesian Network

    Yan Jing, Hori Takayuki, Ohya Jun

      12 ( 3 ) 169 - 170  2013.08

    CiNii

  • H-013 Study of Extracting Moving Objects from Dynamic Kinect Images and Segmenting the Moving and Still Objects' Reconstructed 3D Shape : Proposing a combined method Probabilistic Boundary Edge Map with supervoxel

    Tatematsu Naotomo, Ohya Jun, Davis Larry

      12 ( 3 ) 119 - 120  2013.08

    CiNii

  • H-008 Improving a Bayesian Network Based Recognition of Spontaneous Facial Expressions of a Person who Watches Web News : Utilizing Image Features for Blinks

    Xu Chao, Ohya Jun

      12 ( 3 ) 109 - 110  2013.08

     View Summary

    Recently reading news through web news media is becoming popular. Most of web news is delivered together with a comment system, which asks the user to rate items such as "the news let you think", and "boring". However, many users do not like manual operations for rating; therefore, not many users actually input the rates. One solution for this issue is to utilize results of recognizing facial expressions from the video sequence acquired by the camera that observes the user. So far, there are very many works on recognizing six fundamental expressions such as sad, surprise and happy, for example, by HMM (Hidden Markov Models). However, not many works dealt with recognizing expressions that could appear when the user watches web news; main difficulty in recognizing these expressions could be caused by the fact that these expressions are spontaneously generated. It is difficult to recognize the spontaneous expressions, which come with tiny movements hard to be detected. This paper proposes a method that aims at recognizing spontaneous expressions.

    CiNii

  • H-007 Visual-based Human Gesture Recognition Using Multi-modal Features

    Luo Dan, Ohya Jun

      12 ( 3 ) 107 - 108  2013.08

    CiNii

  • H-038 Fundamental Study of Recognizing the Surgeon's Action during Suture Surgery from the Video Sequence

    Li Ye, Ohya Jun, Chiba Toshio, Xu Rong, Yamashita Hiromasa

      12 ( 3 ) 179 - 180  2013.08

    CiNii

  • H-012 Study of tongue detection and feature extraction for Tongue Diagnosis

    Tian Ming, Luo Dan, Ohya Jun

      12 ( 3 ) 117 - 118  2013.08

    CiNii

  • H-019 Fundamental Study of Particle Filter based Motorbike Tracking for a Violator Detection System

    Do Khanh N., Ohya Jun

      12 ( 3 ) 135 - 136  2013.08

    CiNii

  • H-047 Endoscope based gastric 3D reconstruction and surface mosaic

    Duan Mengyao, Xu Rong, Ohya Jun

      12 ( 3 ) 197 - 198  2013.08

    CiNii

  • K-055 Fundamental Study of Lip-Reading Using Record of Optical Flow Directions and Support Vector Machine

    Takahashi Shohei, Ohya Jun

      12 ( 3 ) 681 - 682  2013.08

    CiNii

  • E-024 Research of Automatic Composition Based on Viewer's Gaze Behavior on Abstract Painting : Fundamental Study of a Method of Generating Tonal Music from the Pattern of Gaze Behavior

    Ogusu Tatsuya, Ohya Jun, Kurumisawa Jun

      12 ( 2 ) 237 - 238  2013.08

    CiNii

  • Study of Robust Facial Feature Localization by Modified Active Shape Model

    Li Yan, Luo Dan, Jun Ohya

    Technical report of IEICE. PRMU   112 ( 495 ) 129 - 134  2013.03

     View Summary

    This paper proposes a modified Active Shape Model (ASM), which is featured by the following two modules: (1)A flexible method that automatically generates training set and does not require time-consuming manual operations, which are required by the conventional ASM; (2)Two-step fitting that consists of the initial global (rough)fitting for the entire face and the subsequent local (precise)fitting for each facial component, where both fittings are based on ASM. Experiments demonstrate that more accurate location results were obtained by the proposed method than the conventional ASM.

    CiNii

  • Fundamental Study of Estimating the Light Source Position by Comparing the 3D Reconstruction Obtained from RGB Image with Depth Image

    Hayashida Tomoki, Takahashi Shohei, Okuya Takeshi, Ohya Jun, Kawamura Harumi

    Technical report of IEICE. Multimedia and virtual environment   112 ( 474 ) 145 - 150  2013.03

     View Summary

    This paper proposes a method for estimating the light source direction from RGB image and Depth image (K-D image), which can be acquired by Kinect sensor etc. The proposed method utilizes the property that the 3D reconstructed result, which is obtained by converting the RGB image to gray-level image and then converting that gray-level image to depth image (D-D image) by Khan et al. 's method based on "Deep is dark paradigm", is deformed to the light source direction. The value difference at each pixel of the D-D image and K-D image is obtained, and that pixel's surface normal is computed from the K-D image. In the 2D space defined by the two angle parameters that represent the surface normal's direction, each pixel's difiFerence is added to the cell that corresponds to that pixel's normal's direction. After this process is performed for all the pixels, the light source direction is determined based on the peak in the parameter space. Experiments using some diflferent objects under different light source directions were performed. Promising results were obtained.

    CiNii

  • Research on Recognizing Customers' Behaviors in Shops Based on Bayesian Network

    ZHENG Yao, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   112 ( 474 ) 51 - 56  2013.03

     View Summary

    For supermarkets and stores, marketing surveillance is a very important issue: in particular, customers' hesitations on purchase, which cannot be recorded by POS systems, are extremely important. This paper studies how to classify the following four categories: "purchased without hesitation", "purchased with hesitation", "not purchased without hesitation", and "not purchased with hesitation". Since these behaviors tend to vary depending on persons and situations, Bayesian network is considered to be a useful method for the classification. By manually extracting image features from video sequences that observe a shop's comer, Bayesian network based classification was explored. Consequently, promising experimental results were obtained.

    CiNii

  • Fundamental Study of Judging Sleeper's Turnovers from Depth Image Sequences by Support Vector Machine

    Minezaki Yuta, Ohya Jun

    Technical report of IEICE. Multimedia and virtual environment   112 ( 474 ) 107 - 112  2013.03

     View Summary

    In recent years, care of aged people's turnovers is heavy burdens for helpers; in particular, automatic judgment of turnovers is desired. This paper proposes a method that can judge whether aged people being cared turned over by utilizing a depth image sequence acquired by Kinect sensor that observes the aged people in beds. Specifically, image features are extracted from aged people's bodies in the depth image sequences. and are used for classification by Support Vector Machine. As the image features, this paper uses time series data of "xyz differences between left and right shoulders and between the left and right edges of the waist". and "average depth in each block obtained by partitioning the bed area into w by h blocks" for exploring classification accuracies. Experimental results show that the former and latter features achieve 99.5% and 89.5% classification accuracies, respectively. The fonner gives a better accuracy, but its applicability is limited. while the latter can be applied for more general cases.

    CiNii

  • Research on Detecting Strangers from Multiple Pan-Tilt-Zoom Static Cameras

    MIZOGUCHI Yoshinori, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   112 ( 474 ) 45 - 50  2013.03

     View Summary

    In this paper, we propose an approach that captures a distant person's face image with high resolution by zooming the camera. Like an ordinary security system, cameras observe down from upper positions. The proposed approach extracts human skin color areas and zooms to the estimated position of the face, even if the resolution of the initial face image is so low that the face cannot be recognized. The captured face image is converted to a DCT (Discrete Cosine Transform)based feature vector, and then SVM (Support Vector Machine) judges if the person is a stranger. The proposed approach shares image information obtained by multiple cameras, and limits the number of zooming camera to one based on 3D position of the moving human in the continuous frames. This could achieve that both large area surveillance and face recognition can be performed simultaneously. Experimental results show validity of the proposed approach.

    CiNii

  • D-12-76 Research on Detecting Strangers from Multiple Pan-Tilt-Zoom Static Cameras

    Mizoguchi Yoshinori, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 169 - 169  2013.03

    CiNii

  • D-16-5 A Robust ICP Algorithm for Automatic Registration of 3D Fetal Facial Surface

    Xu Rong, Ohya Jun, Zhang Bo, Sato Yoshinobu, Fujie Masakatsu

    Proceedings of the IEICE General Conference   2013 ( 2 ) 196 - 196  2013.03

    CiNii

  • D-12-72 Fundamental Study of Judging Sleepers' Turnovers from RGBD ImageSequences by Support Vector Machine

    Minezaki Yuta, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 165 - 165  2013.03

    CiNii

  • D-12-80 Temporal Modified-RANSACとSupervoxelべ一スのグラフカットを用いた移動RBGDカメラからの移動物体領域のセグメンテ-ションの検討

    Tatematsu Naotomo, Ohya Jum, Davis Larry's.

    Proceedings of the IEICE General Conference   2013 ( 2 ) 173 - 173  2013.03

    CiNii

  • D-11-20 Fast Computation of Curvatures for Polygon Models Using Normal Vectors

    Okuya Takeshi, Yabuno Ken, Ohya Jun, Takahashi Nobuyuki

    Proceedings of the IEICE General Conference   2013 ( 2 ) 20 - 20  2013.03

    CiNii

  • D-12-12 Bayesian network model construction for Recognizing Customers, Behaviors in Shops

    ZHENG YAO, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 105 - 105  2013.03

    CiNii

  • D-12-11 Study of Recognizing Spontaneous Facial Behaviors Generated by Persons Watching Web News by Combining ASM and Bayesian Network

    Xu Chao, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 104 - 104  2013.03

    CiNii

  • D-12-10 Face Recognition Using Sparse Representation and Moments

    Li Ye, Shu Huazhong, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 103 - 103  2013.03

    CiNii

  • D-12-21 Robust Lip Contour Localization by Modified Active Shape Model

    Yan Li, Dan Luo, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 114 - 114  2013.03

    CiNii

  • D-11-78 A Study on 3D Point Cloud Stitching for SFM Reconstruction

    Duan Mengyao, Xu Rong, Ohya Jun

    Proceedings of the IEICE General Conference   2013 ( 2 ) 78 - 78  2013.03

    CiNii

  • A Study on Illuminant Color Estimation based on Gray World Assumption using High and Low Chroma Gamuts

    川村春美, 米村俊一, 大谷淳, 小島明

    情報処理学会研究報告(CD-ROM)   2012 ( 6 )  2013

    J-GLOBAL

  • A Study of Automatic 3D Fetal Face Detection By Locating Facial Features From 3D Ultrasound Images for Navigating FETO Surgeries

    XU Rong, OHYA Jun, ZHANG Bo, SATO Yoshinobu, Fujie Masakatsu G.

    IEICE technical report.   112 ( 271 ) 59 - 64  2012.10

     View Summary

    With the wide clinical application of 3D ultrasound (US) imaging, automatic location of fetal facial features from us volumes for navigating fetoscopic tracheal occlusion (FETO) surgeries becomes possible, which plays an important role in reducing surgical risk. In this paper, we propose a feature-based method to automatically detect 3D fetal face and accurately locate key facial features without any priori knowledge or training data. The candidates of the key facial features, such as the nose, eyes, nose upper bridge and upper lip are detected by analyzing the mean and Gaussian curvatures of the facial surface. Each feature is gradually identified from the candidates by a boosted cascade filtering scheme based on the spatial relations between each feature. In experiments, an identification rate of 100% is achieved by using 72 3D US images from a test database of 6 fetal faces in the frontal view and any pose within 15° from the frontal view, and the location error 3.18±0.91mm of the detected upper lip for all test data is obtained, which can be tolerated by the FETO surgery.

    CiNii

  • H-020 Study of Extracting Moving Objects from Dynamic Kinect Images and Segmenting the Moving and Still Objects' Reconstructed 3D Shape : Proposing a Fixation-based Segmentation Method Utilizing Probabilistic Boundary Edge Map

    Tatematsu Naotomo, Ohya Jun, Davis Larry

      11 ( 3 ) 161 - 162  2012.09

    CiNii

  • H-046 Fundamental Study of Recognizing Customers' Behaviors in Shops Based on Bayesian Network

    YAO ZHENG, Ohya Jun

      11 ( 3 ) 221 - 222  2012.09

    CiNii

  • H-048 Fundamental Study of Reconstructing Botanical Tree's 3D Models from Kinect's Images and Dynamical Behaviors of Non-leafy Trees

    Do Khanh N., Ohya Jun

      11 ( 3 ) 225 - 226  2012.09

    CiNii

  • H-042 Fundamental Study of Detecting Strangers from Surveillance Camera

    Mizoguchi Yoshinori, Ohya Jun

      11 ( 3 ) 213 - 214  2012.09

    CiNii

  • H-047 A Study of Extracting 3D Facial Feature from Kinect's Image by Integrating ASM and Depth Map

    LI YAN, LUO DAN, OHYA JUN

      11 ( 3 ) 223 - 224  2012.09

    CiNii

  • H-045 Study of Extracting Multi-modal features for Recognizing Sign Language Vocabularies that Combine Facial Expressions and Hand Gestures

    Luo Dan, Ohya Jun

      11 ( 3 ) 219 - 220  2012.09

    CiNii

  • G-014 A Registration Between 3D Ultrasound Images and 3D Fetal Model for Locating a Fetal Mouth in a Fetal Surgical Navigation System

    Xu Rong, Ohya Jun, Sato Yoshinobu, Fujie Masakatsu G.

      11 ( 2 ) 373 - 374  2012.09

    CiNii

  • Object Tracking Using Kinect and its Application in Tracking Dynamical Behaviors of Botanical Tree Leaves

    DO Khanh N., OHYA Jun

    Technical report of IEICE. HIP   111 ( 500 ) 159 - 164  2012.03

     View Summary

    Reconstructing the dynamical behavior of a moving non-rigid object such as smoke, water and botanical tree is a tough task because it is very difficult to track the behavior of such objects due to their complicated structure and behavior. This paper studies on tracking the dynamical behaviors of a botanical tree by particle filter based methods. The proposed method uses particle filter, which is also known as Sequential Monte Carlo method, to estimate the position of the target leaf (leaf to be tracked) in each frame of the video sequences. The motion model predicts the position of the target leaf by generating different hypothesis of the target leaf's positions. The observation model corrects the prediction by evaluating each "particle" based on template matching score between the leaf model (template) and the target leaf. To increase the tracking accuracy and decrease the computation cost, a leaf is only searched for in its region of interest (ROI). The ROI position and size are updated by one frame delay method.

    CiNii

  • A Cone Based Electromagnetic 3D Ultrasound Calibration

    XU Rong, OHYA Jun, ZHANG Bo, SATO Yoshinobu, FUJIE Masakatsu G.

    Technical report of IEICE. HIP   111 ( 500 ) 253 - 258  2012.03

     View Summary

    For a surgical navigation system with swept-volume 3D ultrasound images, it is necessary to estimate the extracted positions of ultrasound voxels relative to a tracking reference by a 3D ultrasoud calibration. In this paper, we propose a cone based 3D ultrasound calibration system by using an electromagnetic tracking device. The proposed phantom is not expensive, easy to construct, easy to scan, and has a simple experiment setup, while the identification of the tips on each cone from a single ultrasound volume is a manual process, the tracking position of each tip is measured by a pen probe, and the calibration matrix is estimated by SVD (Singular Value Decomposition) method. The precision of the system is experimentally evaluated. It turns out that the maximum distance root mean square (RMS) error is 3.32 mm by using 12 cones, which satisfies the requirement of our fetal surgical navigation system: less than 3.85 mm.

    CiNii

  • Analyzing the "●" Communication System's Effect on Users' Self-Disclosures during Tele-Communication Session

    CHEN Li Jen, OHYA Jun, YONEMURA Shunichi, TOKUNAGA Yukio

    Technical report of IEICE. HIP   111 ( 500 ) 123 - 128  2012.03

     View Summary

    This paper analyzes how our Prompter Communication System (PCS), which has single or several black circular images "●" embedded in the background of the text-field canvas to be used for tele-communication between two users, affects telel-communication users' self-disclosures. To the analysis, this research applies the concepts and features of psychological projection techniques and tests. According to these concepts and features, PCS is expected to enhance and encourage the users' self-disclosure, which is essential to developing intimate and deep interpersonal relationship. Communication experiments by using the prototype system, PCS, are conducted, where a Normal Communication System (NCS), which uses only plain text-field canvas without black circles, is compared with PCS. Two subjects are required to make tele-communication using PCS or NCS. PCS/NCS user writes texts to the canvas with/without randomly generated black circles and sends it (message) to the other user. After tele-communication experiments using PCS and NCS, to each message, for coders give a self-disclosure level, which modifies Daibo et al.'s self-disclosure levels. Consequently, it turns out that PCS gives significantly higher frequencies for the high self-disclosure level, which corresponds to the users' inner thoughts and emotions, than NCS.

    CiNii

  • A Cone Based Electromagnetic 3D Ultrasound Calibration

    XU Rong, OHYA Jun, ZHANG Bo, SATO Yoshinobu, FUJIE Masakatsu G.

    Technical report of IEICE. PRMU   111 ( 499 ) 253 - 258  2012.03

     View Summary

    For a surgical navigation system with swept-volume 3D ultrasound images, it is necessary to estimate the extracted positions of ultrasound voxels relative to a tracking reference by a 3D ultrasoud calibration. In this paper, we propose a cone based 3D ultrasound calibration system by using an electromagnetic tracking device. The proposed phantom is not expensive, easy to construct, easy to scan, and has a simple experiment setup, while the identification of the tips on each cone from a single ultrasound volume is a manual process, the tracking position of each tip is measured by a pen probe, and the calibration matrix is estimated by SVD (Singular Value Decomposition) method. The precision of the system is experimentally evaluated. It turns out that the maximum distance root mean square (RMS) error is 3.32 mm by using 12 cones, which satisfies the requirement of our fetal surgical navigation system: less than 3.85 mm.

    CiNii

  • Object Tracking Using Kinect and its Application in Tracking Dynamical Behaviors of Botanical Tree Leaves

    DO Khanh N., OHYA Jun

    Technical report of IEICE. PRMU   111 ( 499 ) 159 - 164  2012.03

     View Summary

    Reconstructing the dynamical behavior of a moving non-rigid object such as smoke, water and botanical tree is a tough task because it is very difficult to track the behavior of such objects due to their complicated structure and behavior. This paper studies on tracking the dynamical behaviors of a botanical tree by particle filter based methods. The proposed method uses particle filter, which is also known as Sequential Monte Carlo method, to estimate the position of the target leaf (leaf to be tracked) in each frame of the video sequences. The motion model predicts the position of the target leaf by generating different hypothesis of the target leaf's positions. The observation model corrects the prediction by evaluating each "particle" based on template matching score between the leaf model (template) and the target leaf. To increase the tracking accuracy and decrease the computation cost, a leaf is only searched for in its region of interest (ROI). The ROI position and size are updated by one frame delay method.

    CiNii

  • Analyzing the "●" Communication System's Effect on Users' Self-Disclosures during Tele-Communication Session

    CHEN Li Jen, OHYA Jun, YONEMURA Shunichi, TOKUNAGA Yukio

    Technical report of IEICE. PRMU   111 ( 499 ) 123 - 128  2012.03

     View Summary

    This paper analyzes how our Prompter Communication System (PCS), which has single or several black circular images "●" embedded in the background of the text-field canvas to be used for tele-communication between two users, affects telel-communication users' self-disclosures. To the analysis, this research applies the concepts and features of psychological projection techniques and tests. According to these concepts and features, PCS is expected to enhance and encourage the users' self-disclosure, which is essential to developing intimate and deep interpersonal relationship. Communication experiments by using the prototype system, PCS, are conducted, where a Normal Communication System (NCS), which uses only plain text-field canvas without black circles, is compared with PCS. Two subjects are required to make tele-communication using PCS or NCS. PCS/NCS user writes texts to the canvas with/without randomly generated black circles and sends it (message) to the other user. After tele-communication experiments using PCS and NCS, to each message, for coders give a self-disclosure level, which modifies Daibo et al.'s self-disclosure levels. Consequently, it turns out that PCS gives significantly higher frequencies for the high self-disclosure level, which corresponds to the users' inner thoughts and emotions, than NCS.

    CiNii

  • Research on Tracking Fetus' Mouth in Endoscope Video Sequence for Supporting Fetal Surgeries

    XIE Tianliang, XU Rong, OHYA Jun, ZHANG Bo, SATO Yoshinobu, FUJIE Masakatsu G.

    Technical report of IEICE. PRMU   111 ( 499 ) 235 - 240  2012.03

     View Summary

    An ordinary treatment for Congential Diaphragmatic Hernia is Fetal Endoscopic Tracheal Occlusion (FTO), by which a balloon is placed in the trachea so that the lung can be restored in a natural fashion. However, this treatment required very dangerous surgery; thereby, a supporting system for the surgery is demanded. In this paper, to achieve a system that can automatically navigate a tool to be inserted into the fetal trachea, a method that can detect the mouth of the fetus from the endoscopic video sequence is explored. More specifically, a method that combines ellipse detection by randomized Hough transform with CamShift is proposed. Experimentally, the validity of the proposed method is confirmed.

    CiNii

  • A-15-6 Hierarchical Structure for Memories of Faces in Analytic Representation Area : Application to Computer-based Caricatured Portrait Generation System

    Enomoto Makoto, Ohya Jun

    Proceedings of the IEICE General Conference   2012   226 - 226  2012.03

    CiNii

  • A-15-8 Fundamental Study of Generating Portraits Using Neural Networks

    TAKANO Akira, OHYA Jun, ENOMOTO Makoto

    Proceedings of the IEICE General Conference   2012   228 - 228  2012.03

    CiNii

  • A-16-5 Analysis of Gaze Behavior of Viewers Who See Abstract Paintings and Proposal of its Application to Music Composition

    Ogusu Tatsuya, Ohya Jun, Kurumisawa Jun

    Proceedings of the IEICE General Conference   2012   254 - 254  2012.03

    CiNii

  • D-12-15 Human gesture recognition using multimodal features

    Luo Dan, Gao Hua, Ekenel Hazim Kemal, Ohya Jun

    Proceedings of the IEICE General Conference   2012 ( 2 ) 109 - 109  2012.03

    CiNii

  • D-12-13 Real Length Measurement by Kinect

    Duan Mengyao, Xie Yingdi, Ohya Jun

    Proceedings of the IEICE General Conference   2012 ( 2 ) 107 - 107  2012.03

    CiNii

  • D-12-48 Study of the Particle Filter Based Method for Tracking the Dynamical Behaviors of Botanical Tree Leaves in Video Sequence

    Do Khanh N., Ohya Jun

    Proceedings of the IEICE General Conference   2012 ( 2 ) 142 - 142  2012.03

    CiNii

  • D-15-18 Study on a Method of Estimating Learners' Subjective Impressions of the Difficulty for e-Learning on Demand

    Ohkawauchi Takaaki, Ohya Jun, Yonemura Shunichi, Tokunaga Yukio

    Proceedings of the IEICE General Conference   2012 ( 1 ) 215 - 215  2012.03

    CiNii

  • Hierarchical Structure for Memories of Faces in Analytic Representation Area, and Application to Caricatured Portrait Generate System Using Computer

    ENOMOTO Makoto, TAKANO Akira, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   111 ( 479 ) 19 - 24  2012.03

     View Summary

    Viewing the process of drawing caricatured portrait by artist myself introspectively, and it to be a computer program called 'Automatic caricatured portrait generate systems'.

    CiNii

  • Study of Example-based Variations in Virtual Face Makeup Styles

    CHO Jieun, PARK Minje, PARK Sang Il, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   111 ( 479 ) 25 - 29  2012.03

     View Summary

    In this paper, we present a new face makeup analysis and synthesis method for a rapid prototyping of various facial makeup styles. By taking an image pair of a person with and without makeup, we first measure the effect of the given makeup. Then, the extracted makeup is represented by using a finite mixture model. Based on this representation, we provide several operations to easily synthesize a variety of new face makeup effects such as transferring, transplanting, exaggeration, and attenuation. We also demonstrate that multiple makeup styles can be effectively blended to generate a new style in our framework.

    CiNii

  • Fundamental Study of Generating Portraits Using Neural Networks : Studies of Neural Network's Structures and Portraits' Physical and Psychological Similarity

    TAKANO Akira, ENOMOTO Makoto, OHYA Jun

    Technical report of IEICE. Multimedia and virtual environment   111 ( 479 ) 13 - 18  2012.03

     View Summary

    This paper studies a system that can automatically generate portraits from face images using multi-layered neural networks, for which image features of the face are the inputs, the portrait drawn by an illustrator are teaching signals, and back-propagation algorithms are performed for learning. This system aims at outputting portraits that do not depend on deformations of the line drawing traced from the face image, and reflect human feeling. First, this paper explores the neural networks' structures and relationships between the learning's parameter values and the output results. Second, similarity between the system's outputs and portraits drawn by an illustrator is evaluated by subjective tests, and then, using the evaluation results, the relationship between physical errors (physical quantity) and subjective scores (psychological quantity) is studied.

    CiNii

  • Study of Music Composition Utilizing Gaze Behaviors of Persons Viewing Abstract Paintings

    Tatsuya Ogusu, Jun Ohya, Jun Kurumisawa

    IPSJ SIG Notes   2012 ( 18 ) 1 - 6  2012.01

     View Summary

    This paper proposes a method that allows music composition based on abstract paintings. Our approach utilizes gaze behaviors of viewers who see abstract paintings, where temporal changes in gaze positions are utilized as temporal information for music. In this paper, we measured the viewers' gaze behaviors when they see different abstract paintings, and found the following things: (1) durations in which the fix one's gaze on some position are classified into "average length" group and "extremely long" group, and (2) the number of gaze's stay in each element depends on the element's size and distance from the center of the painting. By analyzing mapping between color and sound based on synesthesia, the element at which the gaze stays is converted to sound so that a new music composition method for generating tone clusters is shown.

    CiNii

  • e-learningにおける主観的難易度の補足方法に対する基礎的検討

    大川内隆朗, 大谷淳, 米村俊一, 米村俊一, 徳永幸生

    電子情報通信学会大会講演論文集   2012  2012

    J-GLOBAL

  • WEB上で動作する“黒丸”プロンプタを用いたコミュニケーションシステム-物語をベースにした複数ユーザ間コミュニケーションの特性分析-

    小駒ちなみ, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会全国大会講演論文集   74th ( 4 )  2012

    J-GLOBAL

  • 低彩度および高彩度色領域を用いた灰色仮説に基づく照明光色推定法の一提案

    川村春美, 米村俊一, 大谷淳, 小島明

    画像電子学会年次大会予稿集(CD-ROM)   40th  2012

    J-GLOBAL

  • e-learningにおける学習者の行動履歴を利用した講義改善支援システムの開発

    大川内隆朗, 大谷淳, 米村俊一, 徳永幸生

    情報科学技術フォーラム講演論文集   11th  2012

    J-GLOBAL

  • Improving Iterative Randomized Hough Transform for Automatic Detection of Fetal Head from Ultrasound Images

    Xu Rong, Ohya Jun, Zhang Bo, Sato Yoshinobu, Fujie Masakatsu G.

    IEICE technical report. Image engineering   111 ( 349 ) 127 - 132  2011.12

     View Summary

    In this paper, we propose an improved iterative randomized Hough transform (IRHT) method to detect fetal head automatically in ultrasound images. With the update of region of interest (ROI) in the IRHT method, the noise pixels are gradually excluded from the region of interest during iteration process, and the estimation becomes progressively close to the target. In order to enhance the efficiency and stability of this algorithm, we consider introducing the number (N) of pixels on the detected ellipses, and select the ellipse with the maximal number of the pixels on the ellipse as the result for each iteration, which is selected from the top-M peaks in the accumulators of the whole detected ellipse samples. The experiments on fetal ultrasound images demonstrate that the proposed method achieves more robust and accurate results, and has a better performance for fetal head detection than the IRHT method.

    CiNii

  • J-036 Fundamental Study of the Analysis of the Gaze Behavior of the Viewer for Abstract Paintings and its Application to Automatic Music Composition

    Ogusu Tatsuya, Ohya Jun, Kurumisawa Jun

      10 ( 3 ) 617 - 618  2011.09

    CiNii

  • H-064 Studies of the Kalman filer and Particle filter method for tracking the moving botanical tree leaves in video sequence

    Khanh Do N., Ohya Jun

      10 ( 3 ) 247 - 248  2011.09

    CiNii

  • H-014 Study of Extracting Moving Objects from Dynamic Stereo Images and Segmenting the Moving and Still Objects' Reconstructed 3D Shape : Proposing a SIFT Feature Based Method Utilizing Energy Function for Foreground or Background

    Tatematsu Naotomo, Ohya Jun

      10 ( 3 ) 133 - 134  2011.09

    CiNii

  • G-004 A Study of Automated Fetal Head Detection by Pre-processing based on Ultrasound Image Gray Feature and Iterative Randomized Hough Transform

    Xu Rong, Ohya Jun, Zhang Bo, Sato Yoshinobu, Fujie Masakatsu G.

      10 ( 2 ) 529 - 530  2011.09

    CiNii

  • 顔の輪郭映像を用いた会話システムにおける社会的スキルと注視行動に関する考察

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会全国大会講演論文集   73rd ( 4 )  2011

    J-GLOBAL

  • サイバーシアターにおける観客が実人間あるいはアバターである場合の観客効果の相違に関する検討

    ZHANG Yong, 大谷淳, 米村俊一, 徳永幸生

    電子情報通信学会大会講演論文集   2011  2011

    J-GLOBAL

  • 人物同士のコミュニケーションを触発する“○”メディアにおける自己開示レベルの検討

    CHEN Li Jen, 大谷淳, 米村俊一, 徳永幸生

    電子情報通信学会大会講演論文集   2011  2011

    J-GLOBAL

  • サイバーシアターにおける観客効果に関する検討~観客が実人間・アバターの場合の特性の違いに関する検討~

    YONG Zhang, 大谷淳, 米村俊一, 米村俊一, 徳永幸生

    電子情報通信学会技術研究報告   110 ( 457(MVE2010 133-174) )  2011

    J-GLOBAL

  • 灰色仮説に基づく照明光色推定法の提案

    川村春美, 米村俊一, 大谷淳, 松浦宣彦

    画像電子学会年次大会予稿集(CD-ROM)   39th  2011

    J-GLOBAL

  • Study on a Method of Estimating Learners’ Subjective Impressions of the Difficulty By Their Learning Actions for e-Learning on Demand

    大川内隆朗, 大谷淳, 米村俊一, 米村俊一, 徳永幸生

    電子情報通信学会技術研究報告   111 ( 332(ET2011 76-87) )  2011

    J-GLOBAL

  • 球面領域判定を用いた一枚の画像からの光源位置推定に関する検討

    立松直倫, XU Rong, 大谷淳, 川村春美, 米村俊一

    情報科学技術フォーラム講演論文集   10th  2011

    J-GLOBAL

  • 輪郭映像を用いた会話システムにおける顔注視行動の分析

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会全国大会講演論文集   72nd ( 4 )  2010

    J-GLOBAL

  • “・”プロンプタを用いるコミュニケーションシステムにおけるメッセージ表現の分析

    飯田真也, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会全国大会講演論文集   72nd ( 4 )  2010

    J-GLOBAL

  • サイバーシアターにおいて多数の観客アバターが存在する場合の演技者と観客との間の観客効果に関する基礎的検討

    ZHANG Yong, 大谷淳, 米村俊一, 米村俊一

    電子情報通信学会技術研究報告   110 ( 238(MVE2010 52-73) )  2010

    J-GLOBAL

  • 1チャネルの線画映像コミュニケーションにおける対話ストラテジーと対話特性

    米村俊一, 花田研, 徳永幸生, 大谷淳

    ヒューマンインタフェース学会研究報告集(CD-ROM)   12  2010

    J-GLOBAL

  • 創造的思考を用いたコミュニケーション方法-”・”プロンプタ・システムでの実用について-

    CHEN Li Jen, 米村俊一, 大谷淳, 徳永幸生

    電子情報通信学会技術研究報告   109 ( 471(HIP2009 118-210) )  2010

    J-GLOBAL

  • コミュニケーションにおける会話特性に着目した顔構成要素の役割

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会全国大会講演論文集   71st ( 4 )  2009

    J-GLOBAL

  • 自動会話システムにおけるフォントによる感情表現の効果

    野原雄紀, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   71st ( 4 )  2009

    J-GLOBAL

  • Study on the effect of face parts with line drawing in remote conversation system

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    電子情報通信学会技術研究報告   109 ( 29(WIT2009 1-47) )  2009

    J-GLOBAL

  • “黒丸”プロンプタを用いるコミュニケーションシステムにおける表現特性の分析

    飯田真也, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   71st ( 4 )  2009

    J-GLOBAL

  • “○”プロンプタを用いたコミュニケーションシステムの会話特性

    飯田真也, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   70th ( 4 )  2008

    J-GLOBAL

  • 顔の輪郭画像を用いた会話システムにおける会話特性

    吉田亨, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   69th ( 4 )  2007

    J-GLOBAL

  • “○”プロンプタを用いるコミュニケーションシステム

    原田宣之, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   69th ( 4 )  2007

    J-GLOBAL

  • 線画映像による遠隔サポート-触覚インタラクションの展望-

    米村俊一, 吉田亨, 徳永幸生, 大谷淳

    人間工学   43  2007

    J-GLOBAL

  • Behavioral analysis of novice users on visual support.

    米村俊一, 吉田亨, 徳永幸生, 大谷淳

    電子情報通信学会技術研究報告   106 ( 84(HIP2006 1-22) )  2006

    J-GLOBAL

  • 自動会話システムにおける話題切り替えの効果

    河野恭佑, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   68th ( 2 )  2006

    J-GLOBAL

  • Behavioral analysis of novice users on visual support.

    米村俊一, 吉田亨, 徳永幸生, 大谷淳

    電子情報通信学会技術研究報告   106 ( 219(HCS2006 34-40) )  2006

    J-GLOBAL

  • 顔の輪郭画像を用いた映像通信システムにおける会話特性

    吉田亨, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   68th ( 2 )  2006

    J-GLOBAL

  • 顔のエッジ表現を用いたコミュニケーションシステムの会話特性

    大石貴也, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会全国大会講演論文集   67th ( 4 )  2005

    J-GLOBAL

  • Visual Communication System Based on Edge Abstraction

    大石貴也, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会研究報告   2005 ( 9(HI-112) )  2005

    J-GLOBAL

  • Analysis of Movements of Facial Expressions by High Speed Camera

    UCHIDA Hideko, YOTSUKURA Tatsuo, MORISHIMA Shigeo, YAMADA Hiroshi, OHYA Jun, AKAMATSU Shigeru

    Technical report of IEICE. HIP   99 ( 722 ) 1 - 6  2000.03

     View Summary

    The purpose of this study was to examine patterns of facial movements of "posed" (intended) facial expressions and "elicited" (unintended) emotional responding by feature point tracking. We videotaped participants' facial movements of intended and unintended facial expressions of emotion by a high speed camera which allowed us to analyze facial movements very closely in image sequences. The experiment consisted of two parts. First, the participants task was to produce the six expressions (anger, disgust, fear, happiness, sadness, and surprise). In the second part of the experiment, participants were shown a set of film stimuli that elicited each of emotional states (amusemet, anger, disgust, fear, sadness, and surprise). We recorded the participants' facial expressions to the film stimuli.

    CiNii

  • Analysis of Facial Behevior by Using High-Speed Camera

    Yotsukura Tatsuo, Uchida Hideko, Yamada Hiroshi, Morishima Shigeo, Akamatsu Shigeru, Ohya Jun

    Proceedings of the IEICE General Conference   2000   260 - 260  2000.03

    CiNii

  • Stabilization of Multiple Human Tracking Using Non-synchronous Multiple Viewpoint Observations

    YANG Howard Y, UTSUMI Akira, OHYA Jun

    Technical report of IEICE. PRMU   99 ( 450 ) 1 - 7  1999.11

     View Summary

    We propose a method to track multiple human motion using multiple-camera-based non-synchronous observations. Our method employed Kalman-filtering algorithm to integrate non-synchronous observations from multiple viewpoints effectively. Our experimental system consists of three types of processes, discovering node for detecting newly appeared person, tracking node for tracking each target person, and observation node for processing each viewpoint images. In observation node, occlusion occurred among multiple persons cause serious problem to the matching process between the observed feature and tracked model. In the proposed method, not only position but also human features (height and cloth color) are employed for the matching process to enhance the tracking accuracy. Experimental results with real scene images support our system effectiveness.

    CiNii

  • Eye and Lip Detection and Tracking Using Active Camera

    YOTSUKURA Tatsuo, SHIMADA Naoyuki, MORISHIMA Shigeo, OHYA Jun

    Technical report of IEICE. HIP   99 ( 451 ) 31 - 36  1999.11

     View Summary

    We propose a technique to track user's eye and mouth by using two Pan-Tilt-Zoom controllable cameras. Mouth and eye zones are detected by combine the binary images from the camera. The zoom, rotation direction, and capturing rate of the cameras are automatically changed depends on the results of the captured images. By studying the characteristics of the extracted binary images, we are able to track lip and eye movements such as Lip Reading and blink. Experiments using the technique have shown satisfactory result.

    CiNii

  • 多数カメラによる両手手振り認識を用いた仮想空間操作

    内海章, 大谷淳

    ヒューマンインタフェースシンポジウム論文集   1999   227 - 232  1999.10

    J-GLOBAL

  • 非同期多視点画像を用いた人物追跡システムの構築

    内海章, 大谷淳

    電子情報通信学会大会講演論文集   1999   173  1999.08

    J-GLOBAL

  • Human Motion Detection using Multiple Camera System

    MORI Hiroki, UTSUMI Akira, OHYA Jun, YACHIDA Masahiko

    Technical report of IEICE. Multimedia and virtual environment   98 ( 684(MVE98 98-104) ) 41 - 48  1999.03

     View Summary

    We propose a method to detect human motion information using multiple viewpoint system. In vision-based human tracking, self-occlusions and human-human occlusions are a part of the more significant problems. We employ multiple-viewpoint system to reduce these problems. Our system can retrieve both position and orientation of human bodies, face images and hand positions. Experimental results with five cameras are also described.

    CiNii J-GLOBAL

  • Visual Feedback Device for Non-contact Gesture Interaction in Real Environment

    UTSUMI Akira, OHYA Jun

    Proceedings of the IEICE General Conference   1999  1999.03

    CiNii

  • Human Body Posture Estimation from Trinocular Silhouette Images

    Iwasawa Shoichiro, Ohya Jun, Morishima Shigeo

    Proceedings of the IEICE General Conference   1999 ( 2 ) 264 - 264  1999.03

    CiNii

  • 2)P型フーリエ記述子を用いた多数カメラによる手形状認識(マルチメディア情報処理研究会 ネットワーク映像メディア研究会合同)

    内海 章, 大谷 淳

    映像情報メディア学会誌 : 映像情報メディア   52 ( 12 )  1998.12

    CiNii

  • Kalman-Filter-Based Human Tracking using Asynchronous Multiple Cameras

    UTSUMI Akira, OHYA Jun

    Proceedings of the Society Conference of IEICE   1998   274  1998.09

    CiNii J-GLOBAL

  • Real-Time Human Body Posture Estimation from Multiple Images

    Iwasawa Shoichiro, Takematsu Katsuhiro, Ohya Jun, Morishima Shigeo

    Proceedings of the Society Conference of IEICE   1998   308 - 308  1998.09

    CiNii

  • Multiple-Camera-Based Hand Gesture Recognition using P-Type Fourier Descriptor

    Utsumi Akira, Ohya Jun

    ITE Technical Report   22 ( 39(MIP98 56-63/NIM98 82-89) ) 7 - 12  1998.07

     View Summary

    We propose a hand shape recognition system using multiple-viewpoint images. Most conventional hand shape recognition system did not concern about self-occlusion problem due to the hand rotation and a user needs to pay attention to his hand's direction to avoid the problem. We employ multiple-viewpoint images to estimate the pose of a human hand. After a pose estimation, a "best view" for the hand shape recognition is selected based on the estimation result. Hand shape recognition is performed based on the shape representation using P-type Fourier descriptor that is not affected with image scaling and translation. Based on the shape recognition, we developed a system where a user co create virtual graphical scenes interactively. In the system, a user can change the virtual objects' positions, sizes, colors, etc. with hand gestures. This system can be used as a user interface device, replacing glove-type devices and overcoming most of the disadvantages of contact-type devices.

    CiNii J-GLOBAL

  • Multiple-Camera-Based Multiple-Human Tracking using Kalman Filtering

    UTSUMI Akira, MORI Hiroki, OHYA Jun, YACHIDA Masahiko

    Proceedings of the IEICE General Conference   1998 ( 2 ) 345  1998.03

    CiNii J-GLOBAL

  • 多数カメラによる複数人物の位置・姿勢推定手法の検討

    森大樹, 内海章, 大谷淳, 谷内田正彦

    電子情報通信学会技術研究報告   97 ( 500(PRMU97 196-206) ) 23 - 30  1998.01

    J-GLOBAL

  • ATR知能映像通信研究所

    大谷 淳, 間瀬 健二, 宮里 勉

    日本バーチャルリアリティ学会誌 = Journal of the Virtual Reality Society of Japan   2 ( 2 ) 110 - 112  1997.12

    CiNii

  • 5)多数カメラによる人物位置・姿勢推定(〔ヒューマンインフォメーション研究会 ネットワーク映像メディア研究会〕合同)

    森 大樹, 谷内田 正彦, 内海 章, 大谷 淳

    映像情報メディア学会誌 : 映像情報メディア   51 ( 9 )  1997.09

    CiNii

  • 多数カメラによる実時間手振り推定装置を用いた対話的シーン生成

    内海章, くるみ沢順, 大塚尚宏, 大谷淳

    日本バーチャルリアリティ学会大会論文集   2nd   226 - 227  1997.09

    J-GLOBAL

  • Hand shape recognition using best-view selection

    UTSUMI Akira, OHYA Jun

    Proceedings of the Society Conference of IEICE   1997  1997.08

     View Summary

    筆者らは、非装着型のヒューマンインタフェースへの応用を目指して、画像処理による手振り認識の研究を進めており、多数カメラを利用することで安定な手の位置・姿勢推定が行なえることを示した。推定された手の位置・姿勢は、手形状認識の簡素化および安定化のための視点選択に利用できる。今回、推定姿勢に基づいて正面像を選択することで、フーリエ記述子による形状認識を安定化できることを確認したので報告する。

    CiNii

  • Multiple-Camera Based Estimation of Human Position and Orientation

    MORI Hiroki, UTSUMI Akira, OHYA Jun, YACHIDA Masahiko

    Technical report of IEICE. Multimedia and virtual environment   21 ( 33(HIR97 39-59/NIM97 34-54) ) 21 - 26  1997.06

     View Summary

    We describe a system to estimate position and orientation of a human body. In such a system, deformation and occulusion of the human body often cause difficulties in the estimation. Our system employed multiple-camera based method to avoid these difficulty. Skeleton images detected from multiple cameras are utilized for stable estimation of both position and orientation. Furthermore, according to the estimation result, the best camera for the body side detection is selected. We confirmed the stability of those through experiments.

    CiNii J-GLOBAL

  • Multiple-Camera Based Estimation of Human Position and Orientation

    MORI Hiroki, UTSUMI Akira, OHYA Jun, YACHIDA Masahiko

    ITE Technical Report   12 ( 2 ) 139 - 144  1997.05

     View Summary

    We describe a system to estimate position and orientation of a human body. In such a system, deformation and occulusion of the human body often cause difficulties in the estimation. Our system employed multiple-camera based method to avoid these difficulty. Skeleton images detected from multiple cameras are utilized for stable estimation of both posltion and orientation. Furthermore, according to the estimation result, the best camera for the body side detection is selected. We confirmed the stability of those through experiments.

    CiNii J-GLOBAL

  • Study of Real-Time Human Posture Estimation from Thermal Images

    IWASAWA SHOICHIRO, EBIHARA KAZUYUKI, OHYA JUN, MORISHIMA SHIGEO

    Technical report of IEICE. HCS   96 ( 604 ) 37 - 44  1997.03

     View Summary

    This report proposes a new real-time method that estimates the posture of a human from a thermal image acquired by an infrared camera regardless of background and lighting conditions. Distance transformation is performed for the human body area extracted from the thresholded thermal image so that the center of gravity of body is calculated. After the orientation of the upper half of the body is obtained from calculating the moment of inertia, significant points such as the top of the head, the tips of the hands and foot are heuristically located. In addition, the elbow and knee positions are estimated from the siginificant points using a genetic algorithm based learning procedure. This method does not require to be put any type of device on a body and it can be applied for arbitrary person.

    CiNii

  • Real-Time Estimation of Human Body Posture from Thermal Images

    Iwasawa Shoichiro, Ebihara Kazuyuki, Ohya Jun, Morishima Shigeo

    Proceedings of the IEICE General Conference   1997 ( 2 ) 365 - 365  1997.03

     View Summary

    筆者らは既に、熱画像を用いた非接触な人物の姿勢推定手法を提案している。しかし、筆者らの従来法では、上半身の左右への大幅な傾きや、足の広範な動作に対応できず、また単眼のため3次元情報が得られないなどの課題があった。本報告ではより多くの姿勢に対応できるよう、従来の単眼用アルゴリズムを改良する。即ち単眼赤外線カメラから入力される熱画像から獲得される人物領域と、その輪郭の情報に基づいてヒューリスティックに頭頂・手先・足先の各位置を実時間で検出する手法、および遺伝的アルゴリズム (GA) を利用して肘および膝の位置を推定する手法を提案する。さらに、ステレオ視による3次元位置の獲得について検討する。

    CiNii

  • 時系列画像から推定される画素値分布に基づく手領域抽出

    内海章, 大谷淳

    電子情報通信学会大会講演論文集   1997 ( Sogo Pt 7 ) 361  1997.03

    J-GLOBAL

  • Analysis and Synthesis of Facial Expression Based on Three- dimensional Measurement

    Sakaguchi Tatsumi, Morishima Shigeo, Ohya Jun, Kishino Fumio

    Technical report of IEICE. HC     61 - 68  1994

     View Summary

    We′ve been working on the human-machine interface using facial e xpression animation.But the model based facial expression synthesis method that we proposed are not satisfaction.Because this model deformation rule is constructed based on the two dimensional measurement of human face.In this paper,we propose the three dimensional measurement method for facial surface movement and decide the new deformation rule of facial model.In this method, the three dimensional information is reconstructed from the front and side view image.Furthermore we reconsider the quantify of Action Units in FACS and the interpolation of feature points from the measurement result.

    CiNii

  • Time-sequential range image acquisition by stereo integration of thermal images and intensity images.

    大谷淳, 大和淳司, 石井健一郎

    電子情報通信学会大会講演論文集   1992 ( Shunki Pt 7 ) 7.396  1992.03

    J-GLOBAL

  • Human-independent Action Recognition using Hidden Markov Model from Time-Sequential Images.

    大和淳司, 大谷淳, 石井健一郎

    電子情報通信学会大会講演論文集   1992 ( Shunki Pt 7 ) 7.353  1992.03

    J-GLOBAL

  • Time-sequential range image acquisition by stereo thermal images.

    大谷淳, 大和淳司, 石井健一郎

    電子情報通信学会技術研究報告   91 ( 478(PRU91 126-136) ) 25 - 32  1992.02

    J-GLOBAL

  • Recognizing Human Action in Time-Sequential Images using HMM.

    大和淳司, 大谷淳, 石井健一郎

    電子情報通信学会技術研究報告   91 ( 252(PRU91 51-61) ) 59 - 66  1991.09

    J-GLOBAL

  • Human Action Recognition using HMM.

    大和淳司, 大谷淳, 石井健一郎

    情報処理学会全国大会講演論文集   43rd ( 2 ) 2.245-2.246  1991.09

    J-GLOBAL

  • Range image acquisition by stero thermal image.

    大谷淳, 大和淳司, 石井健一郎

    情報処理学会全国大会講演論文集   43rd ( 2 ) 2.443-2.444  1991.09

    J-GLOBAL

▼display all

Awards

  • The Institute of Image Electronics Engineers of Japan, Excellent Paper Award

    2012.06  

  • 15th Telecommunication Advancement Foundation Award (TELECOM System Technology Prize) by The Telecommunication Advancement Foundation (TAF)

    2000  

  • The 57th Outstanding Patent Applicationss

    1998  

Research Projects

  • Analysis of behaviors of non-rigid objects in real space and their real-time reproduction in 3D virtual space

  • Research on the real-time reproduction of real non-rigid objects' behaviors in 3-dimensional virtual space

  • Research on Reproducing Behaviors of Non-rigid Objects Present in Real Space in Virtual 3D Space

  • Research on Recognizing Human Actions from the Images Acquired by Multiple Static or Dynamic Cameras

  • Study of Analyzing Dynamical Behaviors and 3D Reconstruction from Video Sequences

  • 動画像処理による非剛体の挙動解析の研究

    科学研究費助成事業(早稲田大学)  科学研究費助成事業(基盤研究(C))

▼display all

Presentations

  • Pedestrian Position Estimation by Integrating AlphaPose Based Human Pose Estimation and PredRNN++ Based Spatiotemporal Series Prediction for Autonomous Driving Systems

    Zhang Zhaowei, Eiji Fukuzawa, Jun Ohya

    Presentation date: 2021.08

  • 協生農法環境におけるRGB画像からの圃場の優勢植生の深層学習を用いる検出方法に関する研究

    征矢 寛汰, 青竹 峻太郎, 小方 博之, 大谷 淳, 大谷 拓也, 高西 淳夫, 舩橋 真俊

    2021年度第49回画像電子学会年次大会講演論文集, S7-3, pp.1-4 

    Presentation date: 2021.06

  • 協生農法環境におけるRGB画像に対するSemantic Segmentationを用いた圃場の被覆状態の認識方法に関する研究

    吉崎 玲奈, 青竹 峻太郎, 小方 博之, 大谷 淳, 大谷 拓也, 高西 淳夫, 舩橋 真俊

    2021年度第49回画像電子学会年次大会講演論文集, S8-2, pp.1-4. 

    Presentation date: 2021.06

  • UAVを用いた災害時の道路画像分割に関する研究

    武藤信太, 大谷 淳

    電子情報通信学会画像工学研究会, vol. 120, no. 390, IE2020-70, pp. 97-102 

    Presentation date: 2021.03

  • 環境モニタリングロボットを用いた夜間森林環境モニタリングにおけるSegNetを用いた環境認識と3次元マップ生成に関する研究

    金子丈朗, 大和淳司, 石井裕之, 大谷 淳, 高西淳夫

    電子情報通信学会画像工学研究会, vol. 120, no. 390, IE2020-69, pp. 91-96 

    Presentation date: 2021.03

  • Virtual Reality環境における初心者向けスキー滑走学習システムに関する研究

    宮川峻一, 瀬尾燦振, 大橋拓未, 後藤悠太, 中新かれん, 小澤 悠, 大谷 淳, 彼末一之, 小方博之

    電子情報通信学会画像工学研究会, vol. 120, no. 390, IE2020-61, pp. 52-57 

    Presentation date: 2021.03

  • Switch State Detection by MSRS and YOLOv4 and Automatic Switch Operation with a Robot Arm by Reinforcement Learning in Virtual Environments

    Li Qi, Jun Ohya, Hiroyuki Ogata

    Presentation date: 2021.03

  • Accuracy of Needle Puncture Under Augmented Reality Displayed on a Head Mounted Display with Position Automatically Adjusted by a 2D Marker

    Satoru Morita, Kazufumi Suzuki, Kenji Endo, Takahiro Yamamoto, Shuji Sakai, Shuhei Fujii, Takeshi Morishita, Ryuhei Maruyama, Jun Ohya

    Presentation date: 2020.11

    Event date:
    2020.11
    -
    2020.12
  • ドローンに搭載したカメラ等により獲得される情報を用いる土砂災害の検出法に関する基礎的検討

    松井 開, 長谷川 克也, 武藤 信太, 大谷 淳

    2020年度第48回画像電子学会年次大会講演論文集, S3-4, pp.1-4. 

    Presentation date: 2020.12

  • Classifying Pedestrian Attention for Pedestrian Vehicle Interaction Based on Human Pose Analysis

    Zhao Wentao, Jun Ohya, Zhang Zelin

    Presentation date: 2020.12

  • How to Build AR/MR Phantom for Interventional Radiology Using Medical Images

    Kenji Endo, Kazufumi Suzuki, Satoru Morita, Shuhei Fujii, Ken Masamune, Jun Ohya, Takahiro Yamamoto, Takeshi Morishita, Ryuhei Maruyama, Shuji Sakai

    RSNA2020(Radiological Society of North America), 106th Scientific Assembly and Annual Meeting 

    Presentation date: 2020.11

  • Efficiency of Using Augmented Reality Technology in Training of Central Venous Access

    Kazufumi Suzuki, Kenji Endo, Satoru Morita, Shuhei Fujii, Ken Masamune, Jun Ohya, Takahiro Yamamoto, Takeshi Morishita, Ryuhei Maruyama, Shuji Sakai

    RSNA2020(Radiological Society of North America), 106th Scientific Assembly and Annual Meeting 

    Presentation date: 2020.11

  • Multi-Mask RCNN による多層化マスクを利用した遮蔽された物体の領域分割手法の検討

    平山 慎, 堀 隆之, 大谷 淳

    2020 年 電子情報通信学会総合大会, 情報・システム講演論文集2, D-12-3, p.36 

    Presentation date: 2020.03

  • Virtual Realityを用いたプルーク滑走スキーの体験のためのスキーシミュレータシステムの構築

    宮川峻一, 瀬尾燦振, 大橋拓未, 後藤悠太, 中新かれん, 小澤悠, 彼末一之, 大谷淳, 小方博之

    日本スキー学会第30回大会公演論文集、pp.28-29 

    Presentation date: 2020.03

  • 腹膜播種の光温熱治療のための熱画像の深層学習を用いた高解像度化に関する検討

    三瓶聡太, 尾原睦月, 瀬尾燦振, 長橋 宏, 正宗 賢, 大谷 淳, 守本祐司, 原田 学

    電子情報通信学会IE研究会,信学技報, vol. 119, no. 456, IE2019-118, pp. 103-108 

    Presentation date: 2020.03

  • RGB-Dカメラ画像を用いた災害対応ロボットによる自律的なドアノブ操作方法に関する検討

    林 正晃, 神田琢也, 宮川和也, 大谷 淳, 小方博之, 橋本健二, 孫 瀟, 松澤貴司, 内藤 博, 高西淳夫

    電子情報通信学会IE研究会,信学技報, vol. 119, no. 456, IE2019-109, pp. 63-67 

    Presentation date: 2020.03

  • 環境モニタリングロボットの自己位置推定法選択のためのRGBカメラ画像に対する深層学習による空認識の検討 ~ GNSSかVisual SLAMの選択 ~

    鈴木泰貴, 金子丈朗, 林 拓哉, 森本純矢, 大和淳司, 石井裕之, 大谷 淳, 高西淳夫

    電子情報通信学会IE研究会,信学技報, vol. 119, no. 456, IE2019-98, pp. 11-16 

    Presentation date: 2020.03

  • 自動運転システムのための深層学習を用いた合成画像から実シーン画像への変換および移動経路計画の生成

    張 澤霖, 大谷 淳

    電子情報通信学会PRMU研究会,信学技報, vol. 119, no. 481, PRMU2019-80, pp. 79-84 

    Presentation date: 2020.03

  • ドローンによる視覚障碍者の誘導支援における深層学習を用いた画像からの歩行位置推定に関する研究

    森下 健史, 長橋 宏, 大谷 淳

    画像電子学会研究会,予稿19-03-13、pp.57 - 61 

    Presentation date: 2020.02

  • 術場カメラ画像を用いる脳腫瘍摘出手術の手術工程認識に関する検討

    平田 啓樹, 大谷 淳, 正宗 賢

    画像電子学会研究会,予稿19-03-11、pp.45 - 50 

    Presentation date: 2020.02

  • Microsoft HoloLensを用いたIVR支援ARナビゲーションの知覚誤差に関する研究

    藤井柊平, 森田 賢, 鈴木一史, 遠藤健二, 森下健史, 長橋 宏, 正宗 賢, 大谷 淳

    電子情報通信学会MI研究会,信学技報, vol. 119, no. 399, MI2019-95, pp. 135-139 

    Presentation date: 2020.01

  • 深層学習を用いた心電図からの大動脈弁狭窄症の識別法の提案とGrad-CAMを用いた分析

    秦 絵里香, 瀬尾燦振, 中山雅文, 岩﨑清隆, 大川内隆朗, 大谷 淳

    電子情報通信学会MI研究会, 信学技報, vol. 119, no. 399, MI2019-87, pp. 97-101 

    Presentation date: 2020.01

  • 環境モニタリングロボットのためのRGBカメラ及び深度センサを用いた路面状況認識 〜 SegNet-BasicとENetの比較および高さと曲率特徴の比較 〜

    林 拓哉, 金子丈朗, 森本純矢, 大和淳司, 石井裕之, 大谷 淳, 高西淳夫

    電子情報通信学会技術報告、PRMU研究会 

    Presentation date: 2019.10

  • ガス管内探査ロボットの視覚機能と深層学習を用いた曲管部の湾曲方向検出に関する研究

    宮岡 寛太, 大谷 淳, 徐 栄, 高西 淳夫, 石井 裕之

    第18回情報科学技術フォーラム講演論文集 

    Presentation date: 2019.09

  • Estimating the Positions of Target Objects Based on Object Recognition by Processing 3D Point Cloud Data

    Qi Li, Jun Ohya, Atsuo Takanishi, Takashi Matsuzawa, Kenji Hashimo

    第18回情報科学技術フォーラム講演論文集 

    Presentation date: 2019.09

  • 災害対応ロボットにおけるRGB-Dカメラ情報を用いたスイッチの位置・姿勢の推定

    神田琢也, 大谷淳, 小方博之, 橋本健二, 内藤博, 松澤貴司, 高西淳夫

    第18回情報科学技術フォーラム講演論文集 

    Presentation date: 2019.09

  • 災害対応ロボットWAREC-1による自律的な階段昇段のための画像処理技術の検討

    宮川和也, 大谷淳, 小方博之, 松澤貴司, 橋本健二, 高西淳夫

    第18回情報科学技術フォーラム講演論文集 

    Presentation date: 2019.09

  • フォークリフトの倉庫内の自動運転のための経路周辺の人物の動作識別に基づく危険回避法に関する検討

    吉川 晴音, 林 拓哉, 渋谷 涼太, 大和 淳司, 大谷 淳

    2019 年度 第47 回画像電子学会年次大会 

    Presentation date: 2019.06

  • バーチャルリアリティ技術を用いた第三者視点スポーツトレーニング

    宮川 峻一, 徐 燦振, 小方 博之, 大谷 淳

    2019 年度 第47 回画像電子学会年次大会 

    Presentation date: 2019.06

  • HOG特徴と線形SVMを用いたノイズが含まれる顔動画像からの口唇の検出と口内領域の認識に関する研究

    黒宮 侃熙, 石川 彦太, 平山 慎, 大谷 淳, 徐 栄, 堀 隆之

    画像電子学会第288回研究会講演予稿 

    Presentation date: 2019.03

  • 顔の動画像解析を用いた食事のペース変動に伴うストレス検出方法の検討

    矢野 翔大, 大谷 淳, 大川内 隆朗, 大和 淳司

    画像電子学会第288回研究会講演予稿 

    Presentation date: 2019.03

  • ガス管内探査ロボットの視覚情報を用いた画像変換と進行距離推定

    杉田 愛, 大谷 淳, 徐 栄, 高西 淳夫, 石井 裕之

    画像電子学会第288回研究会講演予稿 

    Presentation date: 2019.03

  • 災害現場における4肢ロボットの深層強化学習を用いた局所的経路生成手法の検討

    黒澤俊樹, 小方博之, 大川内隆朗, 松澤貴司, 大谷 淳, 橋本健二, 高西淳夫

    電子情報通信学会技術報告、PRMU研究会 

    Presentation date: 2019.02

  • QRコード型マーカによる医療用ARナビゲーション

    丸山竜平, 森田 腎, 遠藤健二, 正宗 賢, 長橋 宏, 大谷 淳

    電子情報通信学会技術報告、MI研究会 

    Presentation date: 2019.01

  • 脳腫瘍摘出手術における術者手元近辺動画像中の手の動作解析に基づく手術状況認識法の検討

    堀込貴央, 藤江宏紀, 平田啓樹, 正宗 賢, 長橋 宏, 大谷 淳

    電子情報通信学会技術報告、MI研究会 

    Presentation date: 2019.01

  • 脳外科開頭手術のためのマーカレスARを用いた3次元医用画像の頭部画像への重畳表示

    西崎雄哉, 正宗 賢, 長橋 宏, 大谷 淳

    電子情報通信学会技術報告、MI研究会 

    Presentation date: 2019.01

  • 深層学習とGrad_CAMを用いた超音波画像中の胎児部位検出と深度推定に関する研究

    石川彦太, 徐 栄, 大谷 淳, 岩田浩康

    電子情報通信学会技術報告、MI研究会 

    Presentation date: 2019.01

  • 手術具の検出および追跡による脳腫瘍摘出手術の工程認識法におけるData Augmentationのパラメータ値の検討

    藤江宏紀, 平田啓樹, 堀込貴央, 長橋 宏, 大谷 淳, 田村 学, 正宗 賢, 村垣善浩

    電子情報通信学会技術報告、MI研究会 

    Presentation date: 2019.01

  • 動画像解析による快不快推定の研究

    矢野翔大, 大谷淳, 大川内隆朗, 大和淳司

    第31回人間情報学会予稿集 

    Presentation date: 2018.12

  • 災害対応ロボットにおけるデプスカメラ情報に基づく対象物体の把持位置・方向の自動的な推定

    西川敬士, 大谷 淳, 小方博之, 高西淳夫, 松澤貴司, 橋本健二

    FIT2018 (第17回情報科学技術フォーラム) 

    Presentation date: 2018.09

  • 疾走動作の段階的改善を目的とした教師なし学習による類似動作検出法の検討

    徐 燦振, 佐羽内雅人, 小方博之, 大谷 淳

    FIT2018 (第17回情報科学技術フォーラム) 

    Presentation date: 2018.09

  • 森林環境モニタリングロボットにおける SLAM と深層学習を用いた領域推定に基づく経路計画法の検討

    森本 純矢, 林 拓哉, 藤本 裕之, 石井 裕之, 大谷 淳, 大和 淳司, 高西 淳夫

    2018年度 第46回画像電子学会年次大会 

    Presentation date: 2018.06

  • 脚型ロボットの梯子昇降のための 3 次元点群データ解析による桟の自動的な位置推定に関する検討

    宮川 和也, 西川 敬士, 大谷 淳, 橋本 健二, 高西淳夫

    2018年度 第46回画像電子学会年次大会 

    Presentation date: 2018.06

  • Hand Pose Estimation from Single Depth Images with 3D Convolutional Neural Network

    Zelin Zhang, Zhao Wang, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2018.01

  • 脳動脈瘤の有無と脳動脈血管形状特徴との関係の解析に関する研究

    菊地功成, 八木高伸, 徐 栄, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2017.11

  • ドローンの墜落時点推定に有効なカメラ・センサデータ処理法の基礎的検討

    秦野 勇作, 陳 怡, 大谷 淳

    2017年度 第45回画像電子学会年次大会 

    Presentation date: 2017.06

  • ガス管内探査ロボットの視覚情報を用いた進行距離推定方法に関する研究

    杉田 愛, 徐 栄, 石井 裕之, 大谷 淳, 高西 淳夫

    2017年度 第45回画像電子学会年次大会 

    Presentation date: 2017.06

  • 似顔絵自動生成法のための深層学習による顔画像中の髪領域抽出の検討

    田中 雄也, 大谷 淳, 川村 春美

    2017年度 第45回画像電子学会年次大会 

    Presentation date: 2017.06

  • 環境モニタリングロボットの視覚情報処理を用いた森林における起点への帰還法の検討

    藤本 裕之, 石井 裕之, 大谷 淳, 大和 淳司, 高西 淳夫

    2017年度 第45回画像電子学会年次大会 

    Presentation date: 2017.06

  • RGB-D動画像処理を用いた要介護者の食事状況認識に関する検討

    矢野 翔大, 藤本 篤人, 大川内 隆朗, 大谷 淳

    2017年度 第45回画像電子学会年次大会 

    Presentation date: 2017.06

  • 顔画像中の瞼の応答性を表現する制御モデルによる目の開閉度の連続的な検出法の検討

    徐 燦振, 島 聡志, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2017.03

  • 透視変換を用いたギター演奏時のネックの動画像における追跡法の検討

    本田 浩暉, 王 ?, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2017.03

  • 脳動脈瘤の有無の識別に有効な MRA 画像中の脳血管形状特徴の検討

    齋藤 健太, 菊地 功成, 八木 高伸, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2017.03

  • Support Vector Regressionに基づく3次元動画像処理による人物の動作評価法の検討

    前田尚俊, 王 しょう, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2017.03

  • 縫合手術を構成する6つの段階の動画像中の手術者の手の動きに基づき予測する方法

    坪田実和, 李 イエ, 大谷 淳

    電子情報通信学会研究技術報告 

    Presentation date: 2017.02

  • 違和感の無いアニメーション画像生成のための背景と前景のトーンカーブを用いた自動的な融合法の検討

    江守海渡, 大谷淳, 楜沢順, 川村春美

    FIT2016 (第15回情報科学技術フォーラム) 

    Presentation date: 2016.09

  • Road Vanishing Point Detection by Multi-stage Convolutional Neural Network

    Yan LI, Ye LI, Jun OHYA

    the Media Computing Conference, IIEEJ 

    Presentation date: 2016.06

  • MRA 画像を用いた脳動脈の構造復元のための血流速度依存性と血管内外の輝度値の検討

    菊地功成, 立松直倫, 八木高伸, 大谷 淳, 張澤霖

    画像電子学会研究会 

    Presentation date: 2016.03

  • 縫合手術を構成する5つの段階の動画像からの識別法の検討

    三浦朋也, 李イエ, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2016.03

  • 3次元動画像から抽出される人物の部分骨格モデルを利用する人物の行動認識法の検討

    藤本篤人, 立松直倫, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2016.03

  • SLAM とレーザープロジェクションによる飛翔体ロボットの周囲環境との幾何学的な関係の獲得法の検討

    秦野勇作, 立松直倫, 大谷 淳

    画像電子学会研究会 

    Presentation date: 2016.03

  • Computer Vision Technologies Towards the Actualization of Robotic Scrub Nurses

    Ye Li, Jun Ohya  [Invited]

    日本画像学会誌 

    Presentation date: 2016.02

  • 地域ビジネスとロボット技術

    大谷 淳, 大野邦夫

    画像電子学会 デジタルサイネージとインタラクション(DSG)研究会 第6回ワークショップ 

    Presentation date: 2015.11

  • 抽象絵画における鑑賞者の注視情報を利用した音楽生成システム;抽象絵画の特徴が鑑賞者の抽象的な発想の変化に及ぼす影響

    小楠竜也, 大谷 淳, 楜沢 順, 米村俊一

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • ニューラルネットワークを利用する顔画像からの似顔絵自動生成の検討 -主観評価に基づく比較検討

    郭 帥, 榎本 誠, 川村春美, 大谷 淳

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • 少数特徴点追跡に基づく動画像からの樹木全体の動き追跡の検討

    沈 渊超, 大谷 淳

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • A Method for Tracking Guitar Neck and Fingertips: Necking Tracking Robust against Occlusions Based on Geometry Analysis and Fingertips Tracking Based on Temporal Probability Map

    Zhao Wang, Jun Ohya

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • 超音波画像中の胎児の組織抽出のためのSuperVoxelに基づくセグメンテーション法の検討

    王 元晨, 立松直倫, 大谷 淳, 徐 栄, 石川牧子, 岩田浩康

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • SLAMに基づく片づけロボットによる物体の力学的操作結果の確認方法に関する検討

    高 斯, 立松直倫, 大谷 淳

    FIT2015 (第14回情報科学技術フォーラム) 

    Presentation date: 2015.09

  • 少数特徴点追跡に基づく動画像からの樹木全体の動き追跡の基礎的検討

    シン エンチョウ, 大谷 淳

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • MRI を用いた内頚動脈再構築に関する検討-MRA 画像と血流速度分布のシミュレーションの比較・考察-

    黄 キエン, 八木 高信, 立松 直倫, 大谷 淳, 菊地 功成

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • ニューラルネットワークを用いた顔画像からの似顔絵自動生成 -輪郭と目の描画方法の改良についての検討

    郭 帥, 榎本 誠, 川村 春美, 大谷 淳

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • 絵画の”美しさ”についての基礎的な検討と考察 - 葛飾北斎「富嶽三十六景」への複数被験者による基本図形のあてはめ結果の分析―

    能條 由佳, 川村 春美, 楜沢 順, 大谷 淳

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • Study of Early Recognition and Progress Level Calculation of Surgeon's Action during Suture Surgery

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • SuperVoxel を用いた超音波画像中の胎児に対するセグメンテーションの基礎的検討

    王 元晨, 立松 直倫, 大谷 淳, 徐 栄, 石川 牧子, 岩田 浩康

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • SLAM による自己位置推定と地図生成を用いた片付けロボットの基礎的検討

    高 斯, 立松 直倫, 大谷 淳

    2015年度画像電子学会第43回年次大会 

    Presentation date: 2015.06

  • 人物の関節の三次元データを用いた動作の予測の検討 〜 リー群での相対位置に基づくSkeltonの特徴とMax-Margin Early Event Detectorsを用いた手法 〜

    立松直倫, 大谷 淳, Larry S. Davis

    電子情報通信学会研究技術報告 

    Presentation date: 2015.03

  • ピアノ練習の自動教師システムを目指したRGBD 動画像中の手指追跡の基礎的検討

    邱 東, 大谷 淳

    2015年 電子情報通信学会総合大会 

    Presentation date: 2015.03

  • 抽象絵画における鑑賞者の視線行動を利用した音楽生成システムの研究 −鑑賞者の発話内容の重複と発想の広がりとの関連に関する検討−

    小楠竜也, 大谷淳, 楜沢順, 米村俊一

    2015年 電子情報通信学会総合大会 

    Presentation date: 2015.03

  • Study of a Vision Based Method for Checking the Position of Each Finger of Guitar Players - Towards the Actualization of an Autonomous Guitar Chord Teaching System ?

    Zhao Wang, Jun Ohya

    2015年 電子情報通信学会総合大会 

    Presentation date: 2015.03

  • リー群での人物の関節の相対位置に基づくSkelton の特徴とMax-Margin Early Event Detectors を用いた動作の早期認識の検討

    立松直倫, 大谷 淳, Larry S.Davis

    2015年 電子情報通信学会総合大会 

    Presentation date: 2015.03

  • Fundamental Study of Early Recognition of Surgeon’s Action during Suture Surgery from Video Sequences

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    2015年 電子情報通信学会総合大会 

    Presentation date: 2015.03

  • Recognizing Hand Actions during Suture Surgeries by using sh-SIFT algorithm

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    第32回 日本ロボット学会 学術講演会 

    Presentation date: 2014.09

  • ベイジアンネットワークを用いた店舗内の顧客の購買行動認識法における画像特徴に関する検討

    セイ エン, 堀 隆之, 大谷 淳

    FIT2014(第13 回情報科学技術フォーラム) 

    Presentation date: 2014.09

  • 仮想化粧システムに関する基礎的検討 ― 顔形状の類似度に基づく化粧顔の検索法の検討―

    張 琳, 大谷 淳

    FIT2014(第13 回情報科学技術フォーラム) 

    Presentation date: 2014.09

  • 顔画像からの線画ベースの似顔絵自動生成システムにおける主観的識別に関する検討

    呉 玉珍, 榎本 誠, 大谷 淳

    FIT2014(第13 回情報科学技術フォーラム) 

    Presentation date: 2014.09

  • 抽象絵画における鑑賞者の視線行動を利用した音楽生成システムの研究 ―鑑賞者の発想を膨らませる音楽の自動生成システムに関する予備的検討―

    小楠竜也, 大谷 淳, 楜沢 順, 米村俊一

    FIT2014(第13 回情報科学技術フォーラム) 

    Presentation date: 2014.09

  • 店舗内の消費者の購買行動の動画像処理とベイジアンネットワークによる認識法の検討

    エン セイ, 堀 隆之, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2014.09

  • 顔画像からの似顔絵自動生成システムにおける似顔絵の類似度に関する検討

    呉 玉珍, 榎本 誠, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2014.08

  • 仮想化粧システムに関する基礎的検討

    張 琳, 大谷 淳

    2014年度 画像電子学会第42回年次大会 

    Presentation date: 2014.06

  • 顔画像からの線画ベースの似顔絵の自動生成システムにおける顔パーツ表現に関する検討

    榎本 誠, 呉 玉珍, 大谷 淳

    2014年度 画像電子学会第42回年次大会 

    Presentation date: 2014.06

  • Study of Detecting the Frets and Strings on the Neck of the Guitar from RGBD Images towards the Actualization of an Autonomous Guitar Teaching System

    Zhao WANG, Ye LI, Jing YAN, Jun OHYA

    2014年度 画像電子学会第42回年次大会 

    Presentation date: 2014.06

  • 論文の執筆過程に合わせた文献推薦システムの提案

    大川内 隆朗, 大谷 淳

    2014年度 画像電子学会第42回年次大会 

    Presentation date: 2014.06

  • 抽象絵画における鑑賞者の視線行動を利用した音楽生成システムの予備的検討 〜 提案システムが抽象絵画鑑賞者の発想の多寡に与える影響 〜

    小楠竜也, 大谷 淳, 楜沢 順, 米村俊一

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • ニューラルネットを利用する顔画像からの似顔絵自動生成における最適な特徴点の場所と数の検討

    方 国勇, 榎本 誠, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • 動的輪郭モデルとテクスチャ解析に基づく舌体検出と舌形分類法の検討

    田 明, 羅 丹, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • 移動RGB-Dカメラ画像中からの移動物体の特徴点のクラスタリングの検討 〜 Supervoxelとマルチラベルのグラフカットを用いた手法の提案 〜

    立松直倫, 大谷 淳, Larry S. Davis

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • 顔表情解析に基づく顧客の購買行動認識の基本的検討 〜 ASMによる表情特徴量の取得とベイジアンネットワークのモデルの構築手法 〜

    賈 ロイ, 大谷 淳, 堀 隆之

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • Study of Recognizing Hand Actions from Video Sequences during Suture Surgeries Based on Temporally-Sectioned SIFT and Sliding Window Based Neural Networks

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    電子情報通信学会技術報告 

    Presentation date: 2014.03

  • 複数画像特徴量を用いた読唇システム ― オプテイカルフロー特徴・形状特徴・離散コサイン変換特徴の統合の検討 −

    高橋昌平, 大谷 淳

    情報処理学会研究報告 

    Presentation date: 2014.03

  • Study of Recognizing Surgeon’s Action during Suture Surgery by Using Sectioned-SIFT

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • ニューラルネットを利用する似顔絵自動生成に関する基礎的検討 - 顔特徴点と生成された似顔絵との物理的誤差に基づく最適な特徴点数の検討 ?

    方 国勇, 榎本 誠, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • 舌診における舌体検出と舌形分類法の検討

    田 明, 羅 丹, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • 抽象絵画の鑑賞者の視線行動に基づく音楽生成システムの予備的検討 ― 音楽の有無が抽象絵画鑑賞者の発想の多寡に与える影響 ―

    小楠竜也, 大谷 淳, 楜沢 順, 米村 俊一

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • オプティカルフロー特徴と形状特徴に離散コサイン変換特徴を組み合わせる動画像からの読唇法の検討

    高橋 昌平, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • ニューラルネットワークを用いた顔画像からの線画ベースの似顔絵自動生成システムの検討 ― 鼻の描画法の検討 ―

    呉 玉珍, 榎本 誠, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • 表情解析とベイジアンネットに基づく顧客の購買行動認識に関する検討

    賈 ?維, 大谷 淳, 堀 隆之

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • 画像処理とベイジアンネットワークを用いた店舗内における顧客の全身像からの購買行動認識に関する基礎的検討

    エンセイ, 堀 隆之, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • Supervoxel とマルチラベルのグラフカットを用いた移動物体の特徴点のクラスタリングの検討

    立松 直倫, 大谷 淳, Larry Davis

    電子情報通信学会総合大会 

    Presentation date: 2013.09

  • 抽象絵画における鑑賞者の視線行動を利用した自動作曲に関する研究:視線行動パターンから和声的な音楽の生成方法に関する基礎的検討

    小楠竜也, 大谷 淳, 楜沢 順

    FIT2013 

    Presentation date: 2013.09

  • 舌診における舌体検出と特徴抽出法の検討

    田 明, 羅 丹, 大谷 淳

    FIT2013 

    Presentation date: 2013.09

  • Fundamental Study of Recognizing the Surgeon's Action during Suture Surgery from the Video Sequence

    Ye Li, Jun Ohya, Toshio Chiba, Rong Xu, Hiromasa Yamashita

    FIT2013 

    Presentation date: 2013.09

  • Fundamental Study of Particle Filter based Motorbike Tracking for a Violator Detection System

    N. Khanh Do, Jun Ohya

    FIT2013 

    Presentation date: 2013.09

  • Visual-based Human Gesture Recognition Using Multi-modal Features

    羅 丹, Jun Ohya

    FIT2013 

    Presentation date: 2013.09

  • ニューラルネットを利用する似顔絵自動生成に関する基礎的検討:ASMにより検出される顔特徴と生成された似顔絵の品質との関係の検討

    方 国勇, 榎本 誠, 大谷 淳

    FIT2013 

    Presentation date: 2013.09

  • オプティカルフローの方向特徴履歴とサポートベクトルマシンを用いた読唇の基礎的検討

    高橋昌平, 大谷 淳

    FIT2013 

    Presentation date: 2013.09

  • Improving a Bayesian Network Based Recognition of Spontaneous Facial Expressions of a Person who Watches Web News − Utilizing Image Features for Blinks ?

    Chao Xu, Jun Ohya

    FIT2013 

    Presentation date: 2013.09

  • 移動Kinect画像中からの移動物体の抽出と三次元形状のセグメンテーションに関する検討 −Probabilistic Boundary Edge MapとSupervoxelを組み合わせた手法の提案−

    立松 直倫, 大谷 淳, Larry Davis

    FIT2013 

    Presentation date: 2013.09

  • 画像処理とベイジアンネットワークを用いた店舗内における顧客の全身像からの購買行動認識の基礎的検討

    Jing Yan, 堀 隆之, 大谷 淳

    FIT2013 

    Presentation date: 2013.09

  • Study of Structure from Motion Based 3D Reconstruction and Surface Mosaic of Internal Organs from an Endoscopic Video Sequence

    Mengyao Duan, Rong Xu, Jun Ohya

    FIT2013 

    Presentation date: 2013.09

  • Study of Recognizing Spontaneous Facial Expressions of a Person who Watches Web News Based on ASM and Bayesian Network

    Chao Xu, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2013.09

  • Study of Structure-from-motion Based 3D Reconstruction and Surface Mosaic

    Mengyao Duan, Rong Xu, Jun Ohya

    IEICE Technical Meeting 

    Presentation date: 2013.09

  • Study of Traffic Violator Detection System for Crowded Street −Fundamental Study of Improving Template Based Motorbike Tracking Method−

    Khanh N. Do, Jun Ohya

    2013年度画像電子学会第41回年次大会予稿集 

    Presentation date: 2013.06

  • 抽象絵画を観る人の視線行動を利用した自動作曲に関する研究 −抽象絵画鑑賞時の視線の予測モデルの基礎的検討−

    小楠竜也, 大谷 淳, 楜沢 順

    2013年度画像電子学会第41回年次大会予稿集 

    Presentation date: 2013.06

  • A Robust ICP Algorithm for Automatic Registration of 3D Fetal Facial Surface

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • Temporal Modified-RANSAC とSupervoxel ベースのグラフカットを用いた移動 RBGD カメラからの移動物体領域のセグメンテーションの検討

    立松直倫, 大谷 淳, Larry S. Davis

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • サポートベクターマシンを用いたRGBD 動画像からの寝返り判定に関する基礎的検討

    峰崎 雄大, 大谷 淳

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • Robust Lip Contour Localization by Modified Active Shape Model

    Li Yan, Luo Dan, Jun Ohya

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • 店舗内における消費者の行動認識のためのベイジアンネットワークモデル構築に関する研究

    鄭 曜, 大谷 淳

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • Study of Recognizing Spontaneous Facial Behaviors Generated by Persons Watching Web News by Combining ASM and Bayesian Network

    Chao Xu, Jun Ohya

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • Face Recognition Using Sparse Representation and Moments

    Ye Li, Huazhong Shu, Jun Ohya

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • A Study on 3D Point Cloud Stitching for SFM Reconstruction

    Mengyao Duan, Rong Xu, Jun Ohya

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • RGB 画像からの3次元復元結果とDepth 画像との比較による光源位置推定法に関する基礎的検討

    林田智樹, 高橋昌平, 奥屋武志, 大谷 淳, 川村春美

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • ポリゴンモデルにおける法線ベクトルを用いた曲率の高速計算

    奥屋 武志, 藪野 健, 大谷 淳, 高橋 信之

    2013 年 電子情報通信学会総合大会 

    Presentation date: 2013.03

  • 映像合成のためのRGB画像からの3次元復元結果とDepth画像との比較による光源方向推定法に関する研究

    林田智樹, 高橋昌平, 奥屋武志, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2013.03

  • サポートベクターマシンを用いたDepth動画像からの寝返り判定に関する基礎的検討

    峰崎雄大, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2013.03

  • ベイジアンネットワークを用いた店舗内における消費者の行動認識に関する研究

    鄭 曜, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2013.03

  • 複数の視点固定型パン・チルト・ズームカメラを用いた不審者検出法に関する研究

    溝口慶範, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2013.03

  • Study of Robust Facial Feature Localization by Modified Active Shape Model

    Li Yan, Luo Dan, Jun Ohya

    IEICE Technical Meeting, Pattern Recognition and Machine Understanding 

    Presentation date: 2013.03

  • 映画予告編が映画鑑賞行動に及ぼす影響の物語論に基づく分析

    伊勢田 世山, 高橋 昌平, 奥屋 武志, 大谷 淳

    画像電子学会第264回研究会講演予稿 

    Presentation date: 2013.03

  • 高彩度および低彩度 gamut を用いた灰色仮説に基づく照明光色推定の一提案

    川村春美, 米村俊一, 大谷淳, 小島明

    情報処理学会, 研究報告オーディオビジュアル複合情報処理(AVM) 

    Presentation date: 2013.02

  • A Study of Automatic 3D Fetal Face Detection By Locating Facial Features From 3D Ultrasound Images for Navigating FETO Surgeries

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    IEICE Technical Meeting, Biomedical Engineering and Medical Image Processing 

    Presentation date: 2012.10

  • 移動Kinect 画像中からの移動物体の抽出と三次元形状のセグメンテーションに関する検討

    立松直倫, 大谷淳, Larry Davis

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • A Study of Extracting 3D Facial Feature from Kinect's Image by Integrating ASM and Depth Map

    Li Yan, Luo Dan, Jun Ohya

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • Study of Extracting Multi-modal features for Recognizing Sign Language Vocabularies that Combine Facial Expressions and Hand Gestures

    Luo Dan, Jun Ohya

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • ベイジアンネットワークを用いた店舗内における消費者の行動認識に関する研究

    鄭 曜, 大谷 淳

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • e-learning における学習者の行動履歴を利用した講義改善支援システムの開発

    大川内隆朗, 大谷淳, 米村俊一, 徳永幸生

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • A Registration Between 3D Ultrasound Images and 3D Fetal Model for Locating a Fetal Mouth in a Fetal Surgical Navigation System

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • 防犯用カメラを用いた不審者検出法の基礎的検討

    溝口 慶範, 大谷 淳

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • Fundamental Study of Reconstructing Botanical Tree's 3D Models from Kinect's Images and Dynamical Behaviors of Non-leafy Trees

    Khanh N. Do, Jun Ohya

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • Incorporating Zoom Motion Estimation in Video Compression: A More General Approach for Motion Compensated Predictive Coding

    Morsalin Uz Zoha, Jun Ohya

    FIT2012(第11回情報科学技術フォーラム) 

    Presentation date: 2012.09

  • より効率的な日本式アニメ制作ソフトとその開発に関する研究

    奥屋武志, 大谷 淳, 高橋 信之, 藪野 健

    2012年度 画像電子学会第40回年次大会予稿集 

    Presentation date: 2012.06

  • RGB・距離画像からの照明光源方向の推定に関する検討

    徐 栄, 林田智樹, 大谷 淳, 川村春美

    2012年度 画像電子学会第40回年次大会予稿集 

    Presentation date: 2012.06

  • 低彩度および高彩度色領域を用いた灰色仮説に基づく照明光色推定法の一提案

    川村春美, 米村 俊一, 大谷 淳, 小島 明

    2012年度 画像電子学会第40回年次大会予稿集 

    Presentation date: 2012.06

  • e-learning における主観的難易度の補足方法に対する基礎的検討

    大川内隆朗, 大谷 淳, 米村俊一, 徳永幸生

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • A Study of Rapid Prototyping of Facial Makeup Styles

    Jieun Cho, Minje Park, Sang Il Park, Jun Ohya

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • A Study on 3D Fetal Model Construction with Oral Cavity and Airways

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • FETO手術の自動化のために胎児の口の検出法の基礎的検討

    謝 天亮, 徐 栄, 大谷 淳, 張 博, 佐藤 嘉伸, 藤江 正克

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • 抽象絵画における鑑賞者の視線行動解析と音楽作曲 への応用に関する提案

    小楠 竜也, 大谷 淳, 楜沢 順

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • Human Gesture Recognition Using Multimodal Features

    Luo Dan, Gao Hua, Hazim Kemal Ekenel, Ohya Jun

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • Real Length Measurement by Kinect

    Mengyao Duan, Yingdi Xie, Jun Ohya

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • Study of the Particle Filter Based Method for Tracking the Dynamical Behaviors of Botanical Tree Leaves in Video Sequence

    Khanh N. Do, Jun Ohya

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • ニューラルネットを用いた似顔絵自動生成に関する基礎的検討

    高野 晃, 大谷 淳, 榎本 誠

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • 分析的記憶表象領域における顔記憶の階層的構造 - デフォルメ似顔絵自動生成システムへの応用 -

    榎本 誠, 大谷 淳

    2012年 電子情報通信学会総合大会 

    Presentation date: 2012.03

  • WEB 上で動作する“●”プロンプタを用いたコミュニケーションシステム -物語をベースにした複数ユーザ間コミュニケ-ションの特性分析-

    小駒ちなみ, 米村俊一, 徳永幸生, 杉山 精, 大谷 淳

    情報処理学会第74回全国大会 

    Presentation date: 2012.03

  • Analyzing the "●" Communication System's Effect on Users' Self-Disclosures during Tele-Communication Session

    Li Jen Chen, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    IEICE Technical Report 

    Presentation date: 2012.03

  • A Cone Based Electromagnetic 3D Ultrasound Calibration

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    IEICE Technical Report 

    Presentation date: 2012.03

  • 胎児手術支援システムのための内視鏡画像処理による胎児の口の検出法の基礎的検討

    謝 天亮, 徐 栄, 大谷 淳, 張 博, 佐藤 嘉伸, 藤江 正克

    電子情報通信学会技術研究報告 

    Presentation date: 2012.03

  • Object Tracking Using Kinect and its Application in Tracking Dynamical Behaviors of Botanical Tree Leaves

    Khanh N. Do, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2012.03

  • 化粧サンプルを利用した多様な仮想メイクアップ生成法に関する考察

    智恩, 朴濟, 朴相日, 大谷 淳

    電子情報通信学会技術研究報告 

    Presentation date: 2012.03

  • 分析的記憶表象領域における顔記憶の階層的構造とデフォルメ似顔絵自動生成システムへの応用

    榎本 誠, 高野 晃, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2012.03

  • ニューラルネットを用いた似顔絵自動生成に関する基礎的検討 -ニューラルネットの構成法と物理心理的類似性の検討-

    高野 晃, 榎本 誠, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2012.03

  • 抽象絵画における鑑賞者の視線行動を利用した音楽 作曲の検討

    小楠 竜也, 大谷 淳, 楜沢 順

    情報処理学会研究報告 

    Presentation date: 2012.02

  • オンデマンド型e-learningにおける履修者の学習行動を利用した主観的難易度の把握方法の基礎的検討

    大川内隆朗, 大谷 淳, 米村俊一, 徳永幸生

    電子情報通信学会技術報告 

    Presentation date: 2011.12

  • Improving Iterative Randomized Hough Transform for Automatic Detection of Fetal Head from Ultrasound Images

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    IEICE Technical Report 

    Presentation date: 2011.11

  • A Study of Automated Fetal Head Detection by Pre-processing Based on Ultrasound Image Gray Feature and Iterative Randomized Hough Transform

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    FIT2011(第10回情報科学技術フォーラム) 

    Presentation date: 2011.09

  • 抽象絵画における鑑賞者の視線行動解析と自動音楽作曲への応用に関する基礎的検討

    小楠竜也, 大谷淳, 楜沢順

    FIT2011(第10回情報科学技術フォーラム) 

    Presentation date: 2011.09

  • 球面領域判定を用いた一枚の画像からの光源位置推定に関する検討

    立松 直倫, 徐 栄, 大谷 淳, 川村 春美, 米村 俊一

    FIT2011(第10回情報科学技術フォーラム) 

    Presentation date: 2011.09

  • 移動ステレオカメラ画像中からの移動物体抽出と三次元データのセグメンテーションの検討 -SIFT特徴点を元にした前景・背景のエネルギー関数を組み込んだ手法の提案-

    立松 直倫, 大谷 淳

    FIT2011(第10回情報科学技術フォーラム) 

    Presentation date: 2011.09

  • Studies of the Kalman Filter and Particle filter Method for Tracking the Moving Botanical Tree Leaves in Video Sequence

    Khanh N. Do, Jun Ohya

    FIT2011(第10回情報科学技術フォーラム) 

    Presentation date: 2011.09

  • 灰色仮説に基づく照明光色推定法の提案

    川村春美, 米村俊一, 大谷 淳, 松浦宣彦

    2011年度画像電子学会第39回年次大会 

    Presentation date: 2011.06

  • サイバーシアターにおける観客が実人間あるいはアバターである場合の観客効果の相違に関する検討

    張 勇, 大谷 淳, 米村 俊一, 徳永 幸生

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • A Study of Segmenting Ultrasound Image Based on Bilateral Filtering and Fuzzy Clustering

    Rong Xu, Jun Ohya, Bo Zhang, Yoshinobu Sato, Masakatsu G. Fujie

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • A Study on Edge Feature and its Descriptor

    Yingdi Xie, Jun Ohya

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • 視覚障害者の歩行支援のための動的動画像からのパーティクルフィルタを用いた白線追跡法の検討

    高橋昌平, 大谷 淳

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • 3次元実写画像の彩度を考慮した立体感強調法に関する検討

    古 鎮宇, 大谷 淳, 阿部 信明, 河合 隆史

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • 人物同士のコミュニケーションを触発する“●”メディアにおける自己開示レベルの検討

    Chen Li Jen, 大谷 淳, 米村俊一, 徳永幸生

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • テンソル分解に基づく任意の視点からの人物の動作認識手法の検討

    堀 隆之, 大谷 淳, 楜沢 順

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • Study of Estimating Face Direction and Forefinger Motion for Manipulating a Virtual Object in a Large Screen - Combining Detection of Face Direction and Particle Filter Based Hand and Finger Tracking -

    呉 明翰, 大谷淳

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • Temporal Modified-RANSAC とグラフカットを用いた移動ステレオカメラ画像中からの移動物体抽出と三次元データのセグメンテーションの検討

    立松 直倫, 大谷 淳

    2011年 電子情報通信学会総合大会 

    Presentation date: 2011.03

  • 顔の輪郭画像を用いた会話システムにおける社会的スキルと注視行動に関する考察

    花田 研, 米村俊一, 徳永幸生, 杉山 精, 大谷 淳

    情報処理学会第73回全国大会 

    Presentation date: 2011.03

  • VFT -- A Tree-Based Algorithm that is Robust to Partial-Occlusions and Rotation -

    John Delgaizo, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2011.03

  • 3次元実写画像の立体感強調法に関する検討 〜 彩度と明度を考慮した手法の組み合わせの比較検討 〜

    古 鎮宇, 大谷 淳, 阿部信明, 河合隆史

    電子情報通信学会技術報告 

    Presentation date: 2011.03

  • Study on Manipulating a Virtual Object in a Large Screen Close to the User by Visual Analysis of Small-Scale Movements of Forefinger, Hand and Face

    Ming-Han Wu, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2011.03

  • 移動ステレオカメラ画像中からの移動物体抽出と三次元データのセグメンテーションの検討 〜 Temporal Modified-RANSACと色・構造・事前確率を用いたグラフカットを組み合わせた手法の提案 〜

    立松直倫, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2011.03

  • サイバーシアターにおける観客効果に関する検討 〜 観客が実人間・アバターの場合の特性の違いに関する検討 〜

    張 勇, 大谷 淳, 米村俊一, 徳永幸生

    電子情報通信学会技術報告 

    Presentation date: 2011.03

  • パーティクルフィルタによる白線追跡のためのパラメータの検討 〜 視覚障害者の歩行支援のための人物装着型カメラにより獲得される動画像からの白線認識法の検討 〜

    高橋昌平, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2011.02

  • テンソル分解に基づく多視点映像を用いた人物の動作認識手法の検討

    堀 隆之, 大谷 淳, 楜沢 順

    電子情報通信学会技術報告 

    Presentation date: 2011.02

  • 修正ランザックによる楕円検出

    謝 英弟, 大谷 淳

    画像ラボ 

    Presentation date: 2010.12

  • 意見の対立する場面における会話特性-リアル画像と輪郭画像会話システムにおける注視行動の分析-

    花田 研, 米村俊一, 徳永幸生, 杉山 精, 大谷 淳

    2010年度ヒューマンコミュニケーショングループシンポジウム論文集 

    Presentation date: 2010.12

  • 移動ステレオカメラ画像中からの移動物体の抽出と三次元形状復元およびカメラの姿勢推定を行うTemporal Modified-RANSACに基づく手法の検討

    立松直倫, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2010.12

  • Interactive Guitar: Musical Software That Teaches Guitar Through an Interactive Game Interface and Audio DSP

    John Del Gaizo, Jun Ohya

    IPSJ, SIGMUS 

    Presentation date: 2010.11

  • サイバーシアターにおいて多数の観客アバターが存在する場合の演技者と観客との間の観客効果に関する基礎的検討

    張 勇, 大谷 淳, 米村俊一

    電子情報通信学会技術報告 

    Presentation date: 2010.10

  • Interactive Guitar: Musical Software That Teaches Guitar Through an Interactive Game Interface and Audio DSP

    John Del Gaizo, Jun Ohya

    IPSJ, SIGMUS 

    Presentation date: 2010.10

  • 移動物体を含む環境における移動ステレオカメラの姿勢推定と空間の3次元復元の検討

    立松 直倫, 謝 英弟, 大谷 淳

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • Study of Human Gesture Recognition by Integrating Face and Hand Motion Features

    Luo Dan, Haz?m Kemal Ekenel, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • テンソル分解に基づく多視点映像を用いた人物の動作認識の一検討

    堀 隆之, 大谷 淳, 楜沢 順

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • An Improved KFCM Algorithm in the Segmentation of Brain MRIs

    Rong Xu, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • A System that Allows Human Hand's Short Range Movements to Point a Virtual Object in a Large Screen by Tracking the Hand by Computer Vision Based Approach

    Ming-Han Wu, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • テクスチャマッピング制御による立体感強調の基礎的検討

    古鎮宇, 大谷淳, 河合隆史

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • Studies on the Influences of Economic Responsibility under EPR and Producer’s Adaptive Strategy

    Yiping Zhao, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • Interactive Music: Teaching People Guitar Through Gaming Software

    John Del Gaizo, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • A System for the OMR of Ancient Musical Manuscripts

    Carolina Ramirez, Jun Ohya

    FIT2010(第9 回情報科学技術フォーラム) 

    Presentation date: 2010.09

  • Texture Analysis for Food Recognition

    Do Ngoc Khanh, Jun Ohya, Pishva Davar

    電子情報通信学会技術報告 

    Presentation date: 2010.06

  • 1チャネルの線画映像コミュニケーションにおける対話ストラテジーと対話特性

    米村俊一, 花田 研, 徳永幸生, 大谷 淳

    ヒューマンインタフェース学会研究報告集SIG-CE-01 

    Presentation date: 2010.05

  • Perlin Noise Generated Terrain Raycasting on the GPU

    Sven Forstmann, Jun Ohya

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • C.O.T.O. — A Multimedia Choreography System Inspired by Butoh-fu —

    Jennifer Pierce, Jun Ohya, John Del Gaizo, Ming-Han Wu

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • A Study of Classifying Car Types Using Front and Side Views

    Sebastien Briois, Jun Ohya

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • テンソル分解法に基づくHOG 特徴量を用いた人物の動作認識手法の検討

    堀 隆之, 大谷 淳, 楜沢 順

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • 動物体を含むパン・ティルト映像からのカメラ回転角の推定の検討

    栃折泰史, 江畑 勝, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • 色分け衣服を用いた動画像からの人物姿勢推定法の検討 — パーティクルフィルタにおける"溢れ現象"を利用した人体パーツ追跡法の提案 —

    清水雄太, 大谷淳, 岡崎潤

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • 移動カメラ画像からの手振り動きの検出と認識モデルの構築

    羅 丹, 大谷 淳

    電子情報通信学会総合大会 

    Presentation date: 2010.03

  • "●"プロンプタを用いるコミュニケーションシステムにおけるメッセージ表現の分析

    飯田真也, 米村俊一, 徳永幸生, 杉山 精, 大谷 淳

    情報処理学会第72回全国大会 

    Presentation date: 2010.03

  • 輪郭映像を用いた会話システムにおける顔注視行動の分析

    花田 研, 米村俊一, 徳永幸生, 杉山 精, 大谷 淳

    情報処理学会第72回全国大会 

    Presentation date: 2010.03

  • Policy Design and Producer’s Strategy under Extended Producer Responsibility: Economic Responsibility

    Yiping Zhao, Jun Ohya

    IPSJ SIG Technical Report 

    Presentation date: 2010.03

  • 創造的思考を用いたコミュニケーション方法 - “●”プロンプタ・システムでの実用について -

    Chen Li Jen, 米村俊一, 大谷 淳, 徳永幸生

    電子情報通信学会技術研究報告 

    Presentation date: 2010.03

  • Study of Recognizing Multiple Persons' Complicated Hand Gestures from an Active Image Sequence 〜 Increasing the Number of Persons and Hand Gesture Categories 〜

    Luo Dan, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2010.03

  • Study of Extracting Multiple Independently Moving Objects and Estimating the Egomotion from Active Stereo Sequences

    Yingdi Xie, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2010.03

  • Study of a Car Type classifying Method using an automated Frame segmentation Algorithm -- Study of a specific Representation for Convertibles using local Maxima curvature Function Features ?

    Sebastien Briois, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2010.03

  • パン・ティルト・カメラにより獲得される動画像からのカメラの回転パラメータの推定法に関する検討

    栃折泰史, 大谷 淳, 江畑 勝

    電子情報通信学会技術報告 

    Presentation date: 2010.03

  • 色分け衣服を用いた動画像からの人物姿勢推定のための人体パーツ追跡法の検討 〜 パーティクルフィルタによるオクルージョン判定法の検討 〜

    清水雄太, 大谷 淳, 岡崎 潤

    電子情報通信学会技術報告 

    Presentation date: 2010.03

  • “●”プロンプタを用いるコミュニケーションシステムにおける会話の分析 —GTAによる“●”利用プロセスの分析—

    飯田真也, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    2009年度ヒューマンコミュニケーショングループシンポジウム論文集 

    Presentation date: 2009.12

  • 輪郭映像システムを用いた説得場面における顔構成要素の役割に関する考察

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    2009年度ヒューマンコミュニケーショングループシンポジウム論文集 

    Presentation date: 2009.12

  • カオス(現象)を用いた音楽からリアルタイムに映像を創るアート

    楜沢 順, 下西風澄, 栃折泰史, 立松直倫, 大谷 淳

    第14 回日本バーチャルリアリティ学会大会論文集 

    Presentation date: 2009.09

  • Choreography of the Butoh Object

    Jennifer Pierce, Jun Ohya

    第 14 回日本バーチャルリアリティ学会大会論文集 

    Presentation date: 2009.09

  • Ecological-value Based End-of-life Product Collecting Network Re-design and Optimization

    Yiping Zhao, Jun Ohya

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • 複数カメラを用いたMotion Signatureに基づく人物のトラッキングの検討

    堀 隆之, 大谷 淳, 楜沢 順

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • Proposed Method for OMR of Square Notation Manuscripts

    Carolina Ramirez, Jun Ohya

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • Comparison of several Shape Descriptors in order to perform Car Recognition from a moving Camera

    Sebastien Briois, Xie Yingdi, Jun Ohya

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • 修正RANSACに基づく楕円検出手法の性能評価に関する検討

    張 勇, 謝 英弟, 大谷 淳

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • 色分け衣服を用いた人物の姿勢推定法における人体パーツ追跡法の基礎的検討

    清水雄太, 大谷 淳

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • Automatic shape classification through descriptor matching

    謝 英弟, 大谷 淳

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • 回転運動するカメラにより撮影された動画象からのカメラ回転角の導出法の検討

    栃折泰史, 江畑 勝, 大谷 淳

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • 移動カメラ動画像からの手振り動作認識法の多人数動作への拡張法の検討

    羅 丹, 大谷 淳

    FIT2009(第8回情報科学技術フォーラム) 

    Presentation date: 2009.09

  • ニンフェアール第5回公演 〜息の領域〜

    カミラ・ホイテンガ, 森川栄子, 伊藤美由紀, 大村久美子, 松本祐一, 楜沢 順, 伊庭 崇, 成見 哲, 栃折泰史, 立松直倫, 大谷 淳,他

    サントリー音楽財団推薦コンサート 

    Presentation date: 2009.06

  • 画像処理を用いたサフランめしべ自動採集法の検討

    高野 晃, 小方 博之, 村松 大吾, 大谷 淳

    ロボティクス・メカトロニクス講演会2009 (ROBOMEC 2009)予稿集[No.09-4] 

    Presentation date: 2009.05

  • 顔構成要素の輪郭映像が遠隔対話に与える影響に関する考察

    花田研, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    ヒューマンインタフェース学会研究報告集 

    Presentation date: 2009.05

  • サイバーシアターにおける観客効果に関する基礎的検討

    劉 強, 大谷 淳

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • テンソル分解法に基づく画像処理による個人認証における画像特徴量の比較検討

    堀 隆之, 大谷 淳, 楜沢 順

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • テーブルゲーム時の競技者の顔画像からの視線検出法の検討

    唐 華斌, 大谷 淳

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • A Local Method for Ellipse Detection

    Yingdi Xie, Jun Ohya

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • Creative Thinking Support Communication System with Random Prompter “●” that Allows Hand-drawing

    Li Jen Chen, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • Local Adapaptive Thresholding Study for Early Musical Mananuscripts OMR

    Carolina Ramirez, Jun Ohya

    2009年電子情報通信学会総合大会 

    Presentation date: 2009.03

  • コミュニケーションにおける会話特性に着目した顔構成要素の役割

    花田 研, 徳永幸生, 杉山精, 米村俊一, 大谷 淳

    情報処理学会第71回全国大会 

    Presentation date: 2009.03

  • “●”プロンプタを用いるコミュニケーションシステムにおける表現特性の分析

    飯田真也, 徳永幸生, 杉山精, 米村俊一, 大谷 淳

    情報処理学会第71回全国大会 

    Presentation date: 2009.03

  • 自動会話システムにおけるフォントによる感情表現の効果

    野原雄紀, 徳永幸生, 杉山精, 米村俊一, 大谷 淳

    情報処理学会第71回全国大会 

    Presentation date: 2009.03

  • Stakeholder System and Analysis of Extended Producer Responsibility Based on Game Theory

    Yiping Zhao, Jun Ohya

    情報処理学会研究報告 

    Presentation date: 2009.03

  • サイバーシアターにおける観客の表現法に関する検討

    劉 強, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2009.03

  • 2方向顔画像からのテーブルゲーム競技者の視線検出法に関する検討

    唐 華斌, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2009.03

  • Ellipse Recognition and its Application toward Medical Imaging via Edge Orientation Feature with A Modified RANSAC

    謝 英弟, 大谷 淳

    電子情報通信学会技術報告 

    Presentation date: 2009.03

  • Staff Removal for Square Musical Notation Manuscripts OCR: An Experiment on the Binarization and Staff Detection Stage

    Carolina Ramirez, Jun Ohya

    電子情報通信学会技術報告 

    Presentation date: 2009.01

  • テンソル分解に基づく人物の歩行動作を用いた個人認証

    堀 隆之, 大谷 淳, 楜沢 順

    電子情報通信学会技術報告 

    Presentation date: 2008.12

  • 画像処理による未知人物の行動認識の研究

    堀隆之, 大谷淳

    国際情報通信研究科 公開研究会 

    Presentation date: 2008.10

  • 視覚情報処理の発展と仮想コミュニケーション環境の未来

    大谷 淳

    第38回高柳記念未来技術フォーラム 

    Presentation date: 2008.09

  • The Prototype System of Prompter Communication with Handwriting and Drawing Functions

    Li Jen Chen, Sven Forstmann, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • Application of Synergetic Theory in Municipal Renewable Resource -- Net Values Model of End-of-life Products

    Yiping Zhao, Yingdi Xie, Jun Ohya

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • 顔画像からの顔方向推定のための鼻検出法の比較検討

    唐 華斌, 大谷 淳

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • Study of Detecting Primitive Shapes by Local Optimization Method

    Yingdi Xie, Jun Ohya

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • 複数カメラ画像からの色分け衣服装着人物の3 次元姿勢推定についての検討

    岡崎 潤, 大谷 淳

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • Study of Optimizing the Parameters in the Weighted Histogram Equalization for Stereoscopic Enhancement

    Xiao Ding, Jun Ohya, Takashi Kawai, Nobuaki Abe

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • シルエット特徴量のテンソル分解に基づく動画像から人物の動作認識手法の検討

    堀 隆之, 楜沢 順, 大谷 淳

    FIT2008(第7回情報科学技術フォーラム) 

    Presentation date: 2008.09

  • 複数カメラ画像からの色分け衣服装着人物のオクルージョンが発生する姿勢を含む姿勢推定法の検討

    岡崎潤, 大谷淳

    電子情報通信学会技術報告 

    Presentation date: 2008.09

  • Study of Stereoscopic Constrast Enhancement Based on Local Contrast and Distance Information

    Xiao Ding, Jun Ohya, Takashi Kawai, Nobuaki Abe

    電子情報通信学会技術報告 

    Presentation date: 2008.07

  • 互いに類似した動作のテンソル分解に基づく動画像からの認識法の検討

    堀 隆之, 大谷 淳, 楜沢 順

    電子情報通信学会技術報告 

    Presentation date: 2008.07

  • e-learning における講義ビデオのスキミング手法に関する提案と実装

    大川内隆朗, 大谷淳

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • スポーツ映像において高速に移動するボールを検出する方法の基礎的検討

    劉 強, 謝 英弟, 大谷 淳

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • Study of 3D Contrast Enhancement Based on Weighted Histogram Equalization

    Xiao Ding Jun Ohya, Takashi Kawai, Nobuaki Abe

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 移動カメラ画像からの3次元物体形状復元におけるカメラの移動方向評価尺度に関する検討

    成澤 祐, 大谷 淳, 小方博之

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • Detecting Ellipses from Image by Randomized Hough Transform with Edge Orientation Difference Curve Based Clustering

    Yingdi Xie Jun Ohya

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • Visualizing Large Compressed Volume Data Using a CUDA Compatible GPU

    Sven Forstmann Jun Ohya

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 移動カメラ画像からの人物追従局所座標による手振り軌跡の検出法

    羅丹, 大谷淳

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 色分け衣服装着人物の実写画像に対するMean-Shift アルゴリズムによる色領域分割の検討

    岡崎潤, 大谷淳, Huimin Guo

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 実写動画像からの人物の動作認識のために有効な特徴量の検討

    王偉卿 大谷淳

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 近赤外線画像を用いた人物の鼻の位置からの視線推定法の基礎的検討

    唐 華斌, 大谷 淳, 大川内隆朗

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 顔の特徴を分類化および言語表現可能とした似顔絵システム -デフォルメ似顔絵自動生成システム:ACGS ?

    榎本 誠, 大谷 淳, 楜沢 順

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • A Fundamental Study of “Bi-Conversation” Using “●” Communication Method

    Li Jen Chen, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    2008年電子情報通信学会総合大会 

    Presentation date: 2008.03

  • 自動会話システムにおけるキャラクタの可変表示効果

    野原雄紀, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    情報処理学会第70回全国大会 

    Presentation date: 2008.03

  • ”●“プロンプタを用いたコミュニケーションシステムの会話特性

    飯田真也, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    情報処理学会第70全国大会 

    Presentation date: 2008.03

  • 顔の輪郭画像を用いた会話システムにおける視覚特性

    吉田亨, 米村俊一, 徳永幸生, 杉山精, 大谷淳

    情報処理学会第70全国大会 

    Presentation date: 2008.03

  • A Method for Creative Communication Usin Black-Circled Prompters "●" - Analysis of a Method that Allows Users' Freehand Writing -

    Li Jen Chen, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    電子情報通信学会技術報告 

    Presentation date: 2008.03

  • Multiple Shapes Extraction with Guidance to Random Sampling

    Yingdi Xie, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2008.03

  • Visualizing run-length-encoded volume data on modern GPUs

    Sven Forstmann, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2008.03

  • モンテカルロ法とベイズ推定に基づく動的カメラからの三次元形状復元を利用したカメラの移動方向評価尺度に関する検討

    成澤 祐, 大谷 淳, 小方博之

    電子情報通信学会技術研究報告 

    Presentation date: 2008.03

  • Image Contrast Enhancement Based on Distance Information

    Xiao Ding, Jun Ohya, Takashi Kawai, Nobuaki Abe

    電子情報通信学会技術報告 

    Presentation date: 2008.03

  • Comparative Study of Methods for Recognizing Human Actions from a Real Video Sequence

    Weiqing WANG, Jun OHYA

    電子情報通信学会技術報告 

    Presentation date: 2008.03

  • 移動カメラ動画像からのCondensation Algorithmを用いた手振り認識の検討

    羅丹,大谷淳

    電子情報通信学会技術報告 

    Presentation date: 2008.03

  • 顔の特徴を分類化および言語表現可能とした似顔絵システム - A Caricatured Portrait Generate System Which Can Categorize and Describe Facial Feature -

    榎本誠 大谷淳

    日本顔学会誌 

    Presentation date: 2007.09

  • 自動会話システムにおける背景画像による話題支援

    野原雄紀, 徳永幸生, 杉山精, 米村俊一, 大谷淳

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • 動的カメラ画像からの動物体追跡のためのロボットプラットフォームの検討

    羅丹, 謝英第, 大谷淳

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • 画像処理を用いた店舗内の顧客の商品選定時における態度の認識手法の一考察

    堀 隆之, 楜沢 順, 大谷 淳

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • Study of Foreground-Background Contrast Enhancement in Color Image

    Xiao DING, Jun OHYA

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • カメラの上下動を考慮するモンテカルロ法とベイズ推定に基づく物体の三次元形状復元法の検討

    成澤 祐, 小方博之, 大谷 淳

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • 顔認識における二重符号化理論とデフォルメ似顔絵との関連性について -コンピュータによる似顔絵生成への応用-

    榎本 誠, 大谷 淳, 楜沢 順

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • Low Level Feature Detection based on Modified Ransac

    Yingdi Xie, Jun Ohya

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • VISUALIZATION OF LARGE RLE-ENCODED VOXEL VOLUMES

    Sven Forstmann, Jun Ohya

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • Tensor Decomposition Framework for Recognizing an Unknown Person’s Action from A Video Sequence Using Image Features

    Acep Irawan, Yingdi Xie, Jun Ohya

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • Validation of the Snake Pedal Model applied to Face Segmentation in Monochromatic Images

    Carolina Ramirez, Claudio Perez, Jun Ohya

    FIT2007(第6回情報科学技術フォーラム) 

    Presentation date: 2007.09

  • A Modified RANSAC Based Method for Extracting Primitive Shapes from Image

    Yingdi Xie, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2007.09

  • Study of Different Image Features Used for Recognizing Unknown Person's Activity Based on Tensor Decomposition Analysis

    Acep Irawan, Yingdi Xie, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2007.09

  • Comparative Study of Segmentation in Monochrome Facial Images: Snake Pedals and Template Matching

    Carolina Ramirez, Claudio Perez, Jun Ohya

    IEICE Technical Report 

    Presentation date: 2007.09

  • 線画映像による遠隔サポートシステム

    米村俊一, 吉田亨, 徳永幸生, 大谷淳

    日本人間工学会第48回大会 

    Presentation date: 2007.06

  • 線画によるデフォルメ映像を用いた遠隔サポートの会話特性

    米村俊一, 吉田亨, 徳永幸生, 大谷淳

    電子情報通信学会技術報告 

    Presentation date: 2007.05

  • Implementation of a Video Sequence Based System for Recognizing an Unknown Person’s Gesture by a Tensor Decomposition Based Approach

    Acep Irawan, Xie Yingdi, Jun Ohya

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • 分析的記憶表象領域における顔記憶の階層的構造とそのコンピュータシミュレーション - デフォルメ似顔絵自動生成システムへの応用 -

    榎本 誠, 大谷 淳, 楜沢 順

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • 観察法に基づく画像処理を用いた顧客の行動認識手法の一考察

    堀 隆之, 楜沢 順, 大谷 淳

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • Study of Contrast Enhancement Based on Shadow Detection

    Xiao Ding, Jun Ohya

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • Stability Improvement for Motion Segmentation Based on Active Stereo Cameras

    Yingdi Xie, Jun Ohya

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • Marker-based PC Detection Using Sample Matching

    Qi Zhang, Jun Ohya, Shunichi Yonemura

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • A Study of Communication Prompted by “●” Communication System - Approach on how prompter effects and inspires ideas and thoughts -

    Li Jen Chen, Nobuyuki Harada, Jun Ohya, Shunichi Yonemura, Yukio Tokunaga

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • Procedural Spline-Skeletons for Organic Structures and Adaptive Architecture

    Sven Forstmann, Jun Ohya

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • A Real-time 3D Human Body Tracking and Modeling System

    Huimin Guo, Jun Ohya, Ping Guo

    2007年電子情報通信学会総合大会 

    Presentation date: 2007.03

  • “●”プロンプタを用いるコミュニケーションシステム

    原田宣之, 徳永幸生, 米村俊一, 大谷 淳

    情報処理学会第69回全国大会 

    Presentation date: 2007.03

  • 自動会話システムにおける会話破綻の分析

    野原雄紀, 徳永幸生, 米村俊一, 大谷 淳

    情報処理学会第69回全国大会 

    Presentation date: 2007.03

  • 顔の輪郭映像を用いた会話システムにおける会話特性

    吉田 亨, 徳永幸生, 米村俊一, 大谷 淳

    情報処理学会第69回全国大会 

    Presentation date: 2007.03

  • 陰影領域と非陰影領域に基づく画像コントラスト強調

    丁 皛, 大谷 淳

    SCCC学会研究会報告 

    Presentation date: 2007.03

  • Skeletal Animation by Spline aligned Deformation on the GPU

    Sven Forstmann, Jun Ohya

    電子情報通信学会技術報告, 画像工学研究会 

    Presentation date: 2007.03

  • Study of Image Contrast Enhancement Based on Shadow and Non-shadow Area

    Xiao Ding, Jun Ohya

    電子情報通信学会技術報告, 画像工学研究会 

    Presentation date: 2007.03

  • A Study of Computer Mediated Communication via the “●” Prompter System - Analysis of the affects on the simulation of thought processes and the inspiration of creative ideas

    Li Jen Chen, Nobuyuki Harada, Shunichi Yonemura, Jun Ohya, Yukio Tokunaga

    電子情報通信学会技術報告, ヒューマンコミュニケーション基礎研究会 

    Presentation date: 2007.03

  • Comparative Study of Marker based Detection and Tracking of PC in the Video Sequence Acquired by the Mobile Camera for Visual Call Center Applications

    Qi Zhang, Jun Ohya, Shunichi Yonemura

    電子情報通信学会技術報告(IE) 

    Presentation date: 2007.03

  • Adaptive Optimization to Modified Ransac Based Motion Segmentation

    Yingdi Xie, Jun Ohya

    電子情報通信学会技術報告(IE) 

    Presentation date: 2007.03

  • A modified RANSAC mechanism and its application

    Yingdi Xie, Jun Ohya

    国際情報通信研究科公開研究発表会予稿集 

    Presentation date: 2006.10

  • 顔の特徴を分類化および言語表現可能とした似顔絵システム

    榎本 誠, 大谷 淳

    第11回日本顔学会大会:フォーラム顔学2006、日本顔学会誌 

    Presentation date: 2006.10

  • Independent motion areas extraction from a modified RANSAC mechanism

    Yingdi Xie, Jun Ohya

    電子情報通信学会技術報告(PRMU) 

    Presentation date: 2006.10

  • Optimization of 3D Egomotion Computation in Image Sequence

    Yingdi Xie, Jun Ohya

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • 分析的記憶表象領域における顔記憶の階層的構造とそのコンピュータシミュレーション - デフォルメ似顔絵自動生成システムへの応用 -

    榎本 誠, 大谷 淳, 楜沢 順

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • Study of Color Contrast Enhancement Based on Wavelet

    Ding Xiao, Jun Ohya

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • 多関節弾性物体モデルのための計算コストと使用メモリ削減法の検討

    渡辺隆史, 大谷 淳, 楜沢 順, 徳永幸生

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • Object Detection by Using Marker Assistance

    Qi Zhang, Jun Ohya, Shunichi Yonemura

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • Analysis of Communication Prompted by Black “●”Communication System - Effects of Emotional Emphasis in Comparison with E-mails and Messengers -

    Li Jen Chen, Nobuyuki Harada, Jun Ohya, Yukio Tokunaga, Shunichi Yonemura

    FIT2006(第5回情報科学技術フォーラム) 

    Presentation date: 2006.09

  • Study of Real-time Marker-based PC Detection for Visual Call Center Applications

    Zhang Qi, Ohya Jun, Yonemura Shunichi

    電子情報通信学会技術報告(PRMU) 

    Presentation date: 2006.09

  • Independent motion areas extraction from a modified RANSAC mechanism

    謝 英弟, 大谷 淳

    電子情報通信学会技術報告(PRMU) 

    Presentation date: 2006.09

  • Study of Color Image Enhancement Based on Wavelet Analysis

    Ding Xiao, Jun Ohya

    電子情報通信学会技術報告(PRMU) 

    Presentation date: 2006.09

  • A Modified RANSAC mechanism: Multiple models extraction algorithm

    Yingdi Xie, Jun Ohya

    情報処理学会研究報告,Computer Vision and Image Media 

    Presentation date: 2006.09

  • 輪郭映像を用いたコミュニケーションの視覚特性

    吉田 亨, 徳永幸生, 米村俊一, 大谷 淳

    ヒューマンインタフェースシンポジウム2006 

    Presentation date: 2006.09

  • デフォルメ映像を用いた遠隔サポートのコミュニケーション

    米村俊一, 吉田 亨, 徳永幸生, 大谷 淳

    電子情報通信学会技術報告(HCS) 

    Presentation date: 2006.08

  • デフォルメ映像通信を用いた初心者ユーザの遠隔サポート

    米村俊一, 吉田 亨, 徳永幸生, 大谷 淳

    電子情報通信学会技術報告(HCS) 

    Presentation date: 2006.05

  • 顔の輪郭画像を用いた映像通信システムにおける会話特性

    吉田亨, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会第68回全国大会 

    Presentation date: 2006.03

  • 自動会話システムにおける話題切り替えの効果

    河野恭佑, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会第68回全国大会 

    Presentation date: 2006.03

  • センサーネットワークを用いる消火活動支援システムのための、画像処理による煙の挙動認識法の検討

    野飼雅弘, 大谷

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • 店舗内での顧客の行動を認識するための色情報を用いた物体の存在確認手法の基礎的検討

    堀 隆之, 大谷 淳, 楜沢 順

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • 弾性物体のように振舞うキャラクターモデルの構築法

    渡辺 隆史, 大谷 淳, 楜沢 順, 徳永 幸生

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • 移動カメラを用いるベイズ更新に基づく逐次的3次元形状復元法の検討 - 特徴点の追跡が不要な方法の提案と基礎的検討 -

    成沢 祐, 小方博之, 大谷 淳

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • 顔の特徴を分類化および言語表現可能とした似顔絵システムの提案

    榎本 誠, 大谷 淳, 楜沢 順

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • Study of Motion Segmentation from Dynamic Stereo Images Acquired by Active Stereo Cameras

    Yingdi Xie, Jun Ohya

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • Study of Designing a Prototype System for Exhibiting Artistic Works with Their Creation Processes

    Ton Srichompoo, Jun Ohya, Jun Kurumisawa, Makoto Enomoto

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • A Study of Communication Prompted by the “●” Communication System - Comparing "with Prompters" with "without Prompters" -

    Li Jen Chen, Jun Ohya, Yukio Tokunaga, Shunichi Yonemura

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • Optimise the parameters of spring models so that the Spring models' motions fit well to the real objects' motions

    Denphol Pornchanoknart Jun Ohya

    2006 年電子情報通信学会総合大会 

    Presentation date: 2006.03

  • “ANALYSIS OF HUMAN-TO-HUMAN COMMUNICATIONS PROMPTED BY THE "●" COMMUNICATION SYSTEM - Comparing "with Prompters" and "without Prompters" -

    Li Jen Chen, Jun Ohya, Yukio Tokunaga, Shunichi Yonemura

    電子情報通信学会技術報告 MVE HCGシンポジウム 

    Presentation date: 2006.03

  • Study of the Creation Processes-Based System for Exhibiting Artistic Works

    Ton SRICHOMPOO, Jun OHYA, Jun KURUMISAWA, Makoto ENOMOTO

    電子情報通信学会技術報告 MVE HCGシンポジウム 

    Presentation date: 2006.03

  • 顔の特徴を分類化および言語表現可能とした似顔絵システムの検討

    榎本 誠, 大谷 淳, 楜沢 順

    電子情報通信学会技術報告、MVE(マルチメディア・仮想環境基礎研究会) 

    Presentation date: 2006.03

  • 店舗内における顧客の行動認識方法の基礎的検討-人物の介入するフレームの検出と色情報を用いた人物と物体との接触確認-

    堀隆之, 大谷淳, 楜沢順

    電子情報通信学会、画像工学研究会 

    Presentation date: 2006.03

  • センサーネットワークを利用する消火活動支援システムの検討 - 画像処理を用いた火災状況認識法の検討 -

    野飼 雅弘, 大谷 淳

    電子情報通信学会、パターン認識・メディア理解研究会 

    Presentation date: 2006.03

  • 複数弾性物体を用いた対話操作が可能な弾性物体キャラクターモデルの構築法の提案

    渡辺隆史, 大谷 淳, 楜沢 順, 徳永幸生

    情報処理学会研究報告 

    Presentation date: 2006.02

  • Visualization of Large Caved Terrains

    Sven Forstmann, Jun Ohya

    電子情報通信学会技術研究報告 

    Presentation date: 2006.02

  • Interactive Visualization of Large ISO-Surfaces

    Sven Forstmann, Jun Ohya

    国際情報通信研究科公開研究発表会予稿集 

    Presentation date: 2005.10

  • Comparative Study of Non-rigid Objects' Animations by Spring Models and Boundary Element Method

    Denphol Pornchanoknart, Takafumi Watanabe, Jun Ohya

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • 画像情報を用いたサイバーアクターの動作制御

    Jieun Cho, 渡辺隆史, 大谷 淳

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Study of Black "●" Prompter Communication System

    Li Jen Chen, Jun Ohya, Yukio Tokunaga, Shunichi Yonemura

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • サイバーシアターの空間移動ディレクション装置の提案

    キムヒョンムン, 大谷 淳, 小方博之

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Study of recognizing an unknown person action from motion capture data based on tensor decomposition

    Kalanov Rovshan, Jun Ohya, Hiroyuki Ogata, Hajime Takada

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Interactive Visualization of Large ISO-Surfaces

    Sven Forstmann, 大谷 淳

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Study of Augmented Reality Based Method for Exhibiting Artistic Works with their Creation Processes

    Srichompoo Ton, Jun Ohya, Syunichi Yonemura, Makoto Enomoto

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Image-based Assistance to Fire Extinguishing System with Mote

    Rubaiyat Yasmin, Jun Ohya, Aoki Yoshimitsu

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • 顔認識における二重符号化理論とデフォルメ似顔絵との関連性について - コンピュータによる似顔絵生成への応用 -

    榎本 誠, 大谷 淳, 楜沢 順

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • Motion Detection Based on Active Stereo Camera Array

    Xingdi Xie, Jun Ohya

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • 関節構造を持つ3次元弾性物体モデルの提案

    渡辺隆史, 大谷 淳, 楜沢 順, 徳永幸生

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • 画像処理を用いた店舗内の顧客の行動認識法に関する基礎的検討

    堀 隆之, 大谷 淳, 楜沢 順

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • センサーネットワークと監視カメラを用いた消火活動支援システムの提案

    野飼雅弘, Yasmin Rubaiyat, 大谷 淳

    FIT2005(第4回情報科学技術フォーラム) 

    Presentation date: 2005.09

  • A Proposal on Active Extensible Stereo Camera Array for High Speed Moving Object Detection

    Yingdi Xie, Jun Ohya

    情報処理学会オーディオビジュアル複合情報処理研究会 AVM2005 

    Presentation date: 2005.07

  • Analysis-based Participant Selection Scheme in Workflow Management System

    Ding Xiao, Cao Jian, Jun Ohya

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • A Study on MOTES to Assist Fire Extinguishing System

    Rubaiyat Yasmin, Jun Ohya, Aoki Yoshimitsu

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • A Fundamental Study of Topic Changes Prompted by the "●" Communication System

    Li Jen Chen, Yukio Tokunaga, Shunichi Yonemura, Jun Ohya

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • A Study of Synthesizing New Human Motions from Sampled Motions Using Tensor Decomposition

    Kalanov Rovshan, Jieun Cho, Jun Ohya

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • 因子分解法を用いた屋外環境下での広範囲における強化現実端末の自己位置・姿勢検出精度向上法の基礎的検討

    藤原裕也, 大谷淳, 野飼雅弘

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • 逆運動学による弾性物体の多関節モデルの検討

    渡辺隆史, 大谷淳, 楜沢順, 徳永幸生

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • 色彩画像処理による人物の3 次元姿勢推定に関する検討

    小沼裕一, 姜東完, 大谷淳

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • Shape from Shadingと遺伝的アルゴリズムを用いた水面形状推定法の基礎的検討

    寺島哲平, 大谷淳, 楜沢順, 徳永幸生

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • サイバーアクターの演技のタイミングに関する検討

    曹智恩, 大谷淳

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • サイバーシアターにおけるアヴァター表現に関する検討 - 画像処理による人物姿勢推定結果のニューラルネットによる改良-

    飯田将也, 姜東完, 大谷淳

    2005年電子情報通信学会総合大会 

    Presentation date: 2005.03

  • ソフトエッジとハードエッジに着目したコンピュータによる絵画作成法

    新原雄介, 徳永幸生, 楜沢順, 大谷淳

    情報処理学会第67回全国大会 

    Presentation date: 2005.03

  • 顔のエッジ表現を用いたコミュニケーションシステムの会話特性

    大石貴也, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会第67回全国大会 

    Presentation date: 2005.03

  • A Fundamental Analysis of Topic Changes Prompted by the "●"

    Li Jen Chen, Jun Ohya, Yukio Tokunaga, Tetsuya Kato

    電子情報通信学会技術研究報告,ヒューマンコミュニケーション基礎 

    Presentation date: 2005.03

  • 多関節化による弾性物体のインタラクティブモデルの提案

    渡辺隆史, 大谷淳, 楜沢順, 徳永幸生

    電子情報通信学会技術研究報告,マルチメディア・仮想環境基礎 

    Presentation date: 2005.03

  • ソフトエッジとハードエッジに着目したコンピュータによる絵画制作とその評価

    新原雄介, 徳永幸生, 楜沢順, 大谷淳

    電子情報通信学会技術研究報告,画像工学研究会 

    Presentation date: 2005.03

  • 顔のエッジ表現を用いたコミュニケーションシステム

    大石貴也, 徳永幸生, 米村俊一, 大谷淳

    情報処理学会研究会報告,ヒューマンインタフェース 

    Presentation date: 2005.01

  • Advanced Fire Extinguishing System: An Application of RFID With Ad hoc Networking

    Rubaiyat Yasmin, Jun Ohya, Aoki Yoshimitsu

    第2回QoSワークショップ予稿集 

    Presentation date: 2004.12

  • 熱転写記録を用いた金色銀色表現法とその応用

    徳永幸生, 古田裕樹, 楜沢順, 大谷淳

    パソコンリテラシ 

    Presentation date: 2004.10

  • Studies of the "●" Communication System

    Li Jen Chen, Tetsuya Kato, Yukio Tokunaga, Jun Ohya

    ヒューマンインタフェースシンポジウム2004 

    Presentation date: 2004.10

  • Research on expressing method of audience in a cyber-theater

    Kay Huang, Jun Ohya, Dong Wan Kang

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • LOCALIZATION OF HUMANOID ROBOT FROM MONOCULAR IMAGE SEQUENCE USING TWO LARGE TEMPLATE MATCHING

    Ni Ni Soe, Kenji Suzuki, Shuji Hashimoto, Jun Ohya

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • A Frame Work for Analysis and Synthesis of Actor's Stylistic Action

    Jieun Cho, Jun Ohya

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • 視覚次元互動システム:対話的な展示支援システムに関する提案 Visual-dimension Interact System(VIS)

    恩田淳, Chiu Pei-Yi, 奥智行, 横井真紀, 長幾朗, 大谷淳

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • Fundamentals of RFID and its Application Analysis in Smoky Area

    Rubaiyat Yasmin, Koji Koike, Aoki Yoshimitsu, Jun Ohya, Takagi Mikio

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • 仮想弾性物体の運動の操作手法に関する検討

    渡辺隆史, 大谷淳

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • A Communication System using a black prompter “●”

    Li Jen Chen, Tetsuya Kato, Yukio Tokunaga, Jun Ohya

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • 色彩画像処理と時系列情報による人物の3 次元姿勢推定に関する検討

    小沼裕一, 姜東完, 大谷淳

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • 屋外環境下での広範囲における強化現実端末の自己位置・姿勢検出精度向上法の基礎的検討

    藤原裕也, 大谷淳

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • Shape from Shading と入力画像のアフィン変換を用いた水面の法線ベクトル推定方法の検討

    寺島哲平, 大谷淳

    FIT2004 (第3回情報科学技術フォーラム) 

    Presentation date: 2004.09

  • Environmental Recognition of Humanoid Robot: Image-based Localization and Mapping

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • Study of Computer Vision Based Analysis of the Botanical Tree's Dynamical Behaviors for the Reproduction in Virtual Space

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • Perceptional Interactive Art Museum - Virtual Museum

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 中世テンペラ絵画の色彩復元を重視するディジタル表示法の基礎的検討

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 色彩画像処理による色分け衣服を装着した人物の複雑な姿勢推定法の検討

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 3次元弾性物体の頂点補間を用いた境界要素法による実時間アニメーションモデルの基礎的検討

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • サイバーシアターにおける「観客」の提案

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 形状変形とテクスチャ張り替えを用いた3次元人物顔再構築法

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 赤外発光ダイオードを用いた屋外環境下での強化現実端末の自己位置・姿勢検出精度向上法の基礎的検討

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • 視聴覚相互作用の予備的検討

    電子情報通信学会総合大会 

    Presentation date: 2004.03

  • “●”を用いたコミュニケーションシステムにおける会話条件・利用方法の解析

    情報処理学会全国大会 

    Presentation date: 2004.03

  • 対比刺激を用いた視聴覚相互作用の予備的検討

    電子情報通信学会技術研究報告 

    Presentation date: 2004.03

  • 初期テンペラ絵画のディジタル表示法に関する検討 - 色彩復元と物質感表現保存の試み -

    電子情報通信学会技術報告 

    Presentation date: 2004.03

  • 仮想的再現を目的とした、動画像中の樹木の挙動解析に基づく風の方向と強さの推定法の基礎的検討

    電子情報通信学会技術報告 

    Presentation date: 2004.03

  • 熱転写記録を用いた金色銀色表現法とその応用

    電子情報通信学会技術報告 

    Presentation date: 2004.02

  • 随所アクセス型分散データベースシステムの研究

    大谷淳, 小池公二

    電気通信普及財団、研究調査報告書 

    Presentation date: 2003.12

  • 動的計画法による実時間ステレオ

    電子情報通信学会技術報告 

    Presentation date: 2003.11

  • Comparative study of 3D reconstruction by use of the dynamic stereo cameras with sensors

    FIT(情報科学技術フォーラム)2003 

    Presentation date: 2003.09

  • サイバーシアターディレクティングシステムでの動作の多様性の検討

    FIT(情報科学技術フォーラム)2003 

    Presentation date: 2003.09

  • サイバーシアターにおける演技者の感情による動作表現の多様性に関する基礎的検討

    Jieun Cho, 大谷淳

    電子情報通信学会技術研究報告 

    Presentation date: 2003.09

  • 衝突を伴う仮想弾性物体のリアルタイムアニメーションモデルの提案 - 境界要素法モデルと実写挙動の再現 -

    電子情報通信学会技術研究報告 

    Presentation date: 2003.09

  • “●”を用いたコミュニケーションシステムにおける会話特性の分析

    ヒューマンインタフェースシンポジウム2003 

    Presentation date: 2003.09

  • 強化現実技術を用いた復元遺跡観察装置の開発

    3次元画像コンファレンス2003 

    Presentation date: 2003.07

  • サイバーシアターの研究

    大谷淳, 伊藤隆道

    AMC報告(東京芸術大学芸術情報センター編集・発行) 

    Presentation date: 2003.03

  • 実写動画像と顔面筋モデルによる3次元人物顔画像の構築

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 動的ステレオカメラを用いた3次元復元の検討

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 台本のト書きの分析結果を利用したサイバーシアター・ディレクティング・システム

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 境界要素法による3次元ゼリー状物体の挙動の実時間再現

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 3次元樹木モデルにおける効率的かつリアルな葉のアニメーション法の検討

    神田仁, 大谷淳

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 仮想的再現のためのShape from Shadingを利用した水面形状の推定法の基礎的検討

    寺島哲平, 大谷淳

    電子情報通信学会総合大会 

    Presentation date: 2003.03

  • 効率的かつリアルな3次元樹木モデルのアニメーションの検討

    神田仁, 大谷淳

    電子情報通信学会技術研究報告 

    Presentation date: 2003.02

  • 仮想的再現のためのShape from Shadingを利用した水面形状の推定法の基礎的検討

    寺島哲平, 大谷淳

    電子情報通信学会技術研究報告 

    Presentation date: 2003.02

  • 動画像と顔面筋動作を考慮したモデルを用いた3次元人物顔画像構築法

    電子情報通信学会技術研究報告 

    Presentation date: 2003.02

  • 仮想コミュニケーション環境の研究 画像処理による人物の姿勢の実時間3次元推定

    画像ラボ 

    Presentation date: 2002.11

  • 文化財環境における遍在型デジタルアーカイブ研究 - 非接触型ICチップ/タグを利用した、文化財環境でのユビキタス・アーカイブの実現について -

    小池公二, 平尾大輔, 大谷淳

    情報処理学会、人文科学とコンピュータ研究会 

    Presentation date: 2002.11

  • Field Augmented Reality実現のための絶対位置・方位測定方法の検討

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 3Dテクスチャ画像の材質感解析

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 境界要素法を用いた仮想弾性体のリアルタイムアニメーション法

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 力学的挙動の効率的表現が可能な樹木の三次元モデリング法の検討

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 3方向光源差分画像を用いた古銭画像識別法の検討

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 色彩画像処理による色分け衣服を装着した人物の姿勢推定法の検討

    FIT(情報科学技術フォーラム)2002 

    Presentation date: 2002.09

  • 2枚の実写画像による顔面筋動作を考慮した3次元顔特徴モデルの構築法

    電子情報通信学会技術研究報告 

    Presentation date: 2002.07

  • 文化財保存科学における偏在型でじたるアーカイブの研究 ICチップと近距離無線搭載携帯端末によるユビキタス・アーカイブの実現について

    文化財保存修復学会第24回大会研究発表要旨集 

    Presentation date: 2002.06

  • 物体形状の推定と表面質感に関する一考察

    2002画像電子学会第30回年次大会 予稿集 

    Presentation date: 2002.06

  • 実空間中の流体挙動の仮想的再現に関する基礎的検討

    2002画像電子学会第30回年次大会 予稿集 

    Presentation date: 2002.06

  • コンピュータヴィジョンに基づく人物と物体のインタラクション認識の検討

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • 単眼カメラによるリアルタイム人物顔特徴トラッキング法の検討

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • 不変特徴を用いた古銭画像の識別法の検討

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • JPEG2000を用いた静止画中の物体の質感表現の検討

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • 映りこみを利用した水の再現方法

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • 筋肉動作を考慮した3次元顔モデルの構築と表情生成の応用

    電子情報通信学会総合大会 

    Presentation date: 2002.03

  • 接触・非接触混合型多視点カメラ画像からの人物全身像の姿勢推定法の検討

    平成14年電気学会全国大会 

    Presentation date: 2002.03

  • 2001 IEEE International Conference on Multimedia and Expo (ICEM2001, マルチメディアに関する国際会議と展示)

    電子情報通信学会誌 

    Presentation date: 2002.01

  • 画像処理による人物と物体のインタラクション認識の検討

    情報処理学会第63回(平成13年後期)全国大会 

    Presentation date: 2001.09

  • 実写画像を用いた流体挙動の仮想的再現の検討

    2001年電子情報通信学会情報・システムソサイエティ大会 

    Presentation date: 2001.09

  • 3D顔モデル生成方法の比較検討

    2001年電子情報通信学会情報・システムソサイエティ大会 

    Presentation date: 2001.09

  • 仮想コミュニケーション環境の研究 人物像の画像処理による認識とCG技術による生成

    画像ラボ 

    Presentation date: 2001.06

  • 仮想コミュニケーション環境の研究

    3D映像 

    Presentation date: 2000.12

  • 高速度カメラを用いた顔面動作の分析

    2000年電子情報通信学会総合大会 基礎・境界ヒューマンコミュニケーション基礎 

    Presentation date: 2000.03

  • 高速度カメラを用いた顔面表情の動的変化に関する分析

    第33回知覚コロキウム 

    Presentation date: 2000.03

  • ニューラルネットワークによる実時間人物姿勢推定法に関する一考察

    電子情報通信学会総合大会 情報システム 

    Presentation date: 2000.03

  • 実写映像に基づく枝振りを考慮した樹木のモデリング手法

    電子情報通信学会総合大会 情報システム 

    Presentation date: 2000.03

  • ATRにおける仮想コミュニケーション環境の研究

    電子情報通信学会PRMU研究会 特別講演 PRMU 99-203 

    Presentation date: 2000.01

  • 分身コミュニケーションの提案と検討—ノンバーバル情報の認識を利用した新たなコミュニケーション手段—

    電子情報通信学会PRMU研究会 

    Presentation date: 1999.11

  • 眉間の検出・追跡による頷きと首振りのリアルタイム検出

    電子情報通信学会PRMU研究会 

    Presentation date: 1999.11

  • 非同期多視点画像を用いた複数人物追跡の安定化

    電子情報通信学会PRMU研究会 

    Presentation date: 1999.11

  • アクティブカメラによる視線追跡・自動Lip Reading

    電子情報通信学会ヒューマン情報処理研究会(HIP) 

    Presentation date: 1999.11

  • SIGGRAPH99に参加して

    ATRジャーナル 

    Presentation date: 1999.11

  • 多数カメラによる両手手振り認識を用いた仮想空間操作

    ヒューマンインタフェースシンポジウム'99 

    Presentation date: 1999.10

  • 三眼視による実時間非接触非装着型三次元人物姿勢推定法

    電子情報通信学会PRMU・NC研究会 

    Presentation date: 1999.10

  • 非同期多視点画像を用いた人物追跡システムの構築

    1999年電子情報通信学会基礎・境界ソサイエティ大会講演論文集 

    Presentation date: 1999.09

  • カルマンフィルタを用いた実時間人物姿勢推定法の一考察

    1999年電子情報通信学会基礎・境界ソサイエティ大会講演論文集 

    Presentation date: 1999.09

  • 表情推定のための顔の眉間位置検出法の検討

    1999年電子情報通信学会情報・システム ソサイエティ大会 講演論文集 

    Presentation date: 1999.09

  • 実時間三次元人物姿勢推定法に関する一考察

    日本バーチャルリアリティ学会第4回大会 

    Presentation date: 1999.09

  • 風に揺れる樹木のCGアニメーション

    画像電子学会年次大会 

    Presentation date: 1999.06

▼display all

Specific Research

  • コンピュータ・エージェントの行動生成のための早期環境認識の研究

    2021  

     View Summary

    エージェントが限られた時間の中で認識を行い、その結果から他の主体の行動を決定するシステムの例として、複雑化した手術の工程を早期認識して、手術スタッフあるいは看護師ロボットに伝える方法の実現に資する技術の実現を目指す。具体的には、手術に参加する複数の人物を個別に特定する方法の検討を行った。手術環境において手術着を着用した人物を識別するために、動画像から手術着を着用した人物の骨格を取得し,これに深層学習を適用することにより、人物を特定する方法を提案し、実験により検討した。動画像中の人物の2次元的姿勢をOpenPoseにより推定し、深層学習により3次元姿勢を得る。20名を用いた識別実験の結果、80%の識別率を達成した。

  • 高度手術支援を目指した術中状況認識のデジタル化戦略

    2020  

     View Summary

    本年度は、手術工程の認識を行うために必要な前処理である、手術に参加している人物を個別に切り出す処理について検討を行った。具体的には、まず手術室を俯瞰するように設置されたカメラにより手術室内の医療スタッフを撮影し、カメラの映像より人物の骨格を二次元的に、OpenPoseを用いて抽出する。次に、ディープラーニングにより、抽出された二次元骨格から三次元的に立体化を行う。最後に、この三次元的な骨格を機械学習によって解析し、最終的な人物の識別を行う。7名約50万セットの骨格情報を使用して学習を行った結果、各関節座標の平均絶対誤差16.7mmを達成した。今後は、さらに推定精度の向上を目指すとともに、手術工程の認識の検討を進める。

  • 手術ワークフローのデジタル化のための術中動画像処理・シナリオ自動生成の研究

    2019  

     View Summary

     本研究では,手術室に設置された術場カメラから獲得される動画像から手術工程を認識する方法として,動画像の各フレームにおける執刀医と看護師の関節を特徴点として検出し,Bag of Visual Wordsを用いて分類器を作成し,工程を認識する手法を提案する.特徴点検出には少数のデータにより転移学習を行ったDeepLabCutを用いる.検出された特徴点とそれらを結ぶスケルトンモデルを描画し,検出した特徴点を多次元空間中でクラスタリングし,その結果を利用する工程の分類器を作成した.提案法により未知の動画像から手術工程が認識される.手術室において二人の手術医の模擬手術を撮像することにより獲得された動画像に対して提案手法を適用し、有効性の見通しを得た.

  • 非接触センサからの視覚情報等に基づく人体動作の自動教師システムの研究

    2018  

     View Summary

    近年、スポーツのスキルの上達を目指して、運動者をカメラで撮像し、画像処理と機械学習を用いて人間の運動動作を評価するシステムが注目を集めている。本研究では、少数箇所の改善を段階的に行って動作改善を目指す方法を検討する。まず、教師なし学習を用いて、類似する運動動作を検出するための特徴量を得る方法を検討する。具体的にはAutoencoder を教師無し学習に用いる。次に、得られた生成モデルの情報に対してクラスタ分析を行い、得られるクラスタの特徴を、学習済みモデルにおける入力寄与度を用いて分析する。本研究では短距離走における疾走動作を検討対象として実験を行い、検出されたクラスタにおける疾走動作の傾向が得られる見通しを得た。

  • 視覚情報等に基づく人体動作の自動教師システムの研究

    2018  

     View Summary

    Recently, with the development of computer vision, achieving automatic guitar fingering teaching systems has been attracting a lot of attentions of academic research.  This research proposes a Particle Filter combined with ROI based method for (1) hand extraction, (2) fingertip candidate detection, and (3) fingertip tracking, as follows: (1) we propose an end-to-end CNN framework (some related works also call it as FCN, Fully-connected Network. This network is a dynamic variation of CNN, but we still call it CNN for respecting the origin) with hundreds of labelled hand segmentation images, which could segment the hand area at each frame of guitar playing videos accurately; (2) we combine Template Matching and reversed Hough Transform as the features foraccurately locating fingertip candidates; (3) we apply a temporal-grouping for the candidates based on ROI (region of interest) association to group the same fingertip candidates on consecutive frames and distribute particles in the surrounding area centered at each of the associated fingertip candidates to address the fast movements and self-occlusions of the fingertips.   Experiments are conducted using videos of guitar plays under different conditions. For the hand region segmentation and fingertip tracking, the proposed method outperforms the related works.

  • 動画像処理に基づくギター演奏評価法の研究

    2017  

     View Summary

    ギター演奏者の左手の動きをカメラで撮像することにより獲得される動画像を解析し、運指の良否を自動的に判定する方法の検討を行った。ギター演奏中の手と指の動画像の訓練データを収集する。各指の関節の位置の時系列データに対するDCT(離散コサイン変換)特徴を求める。一方、それぞれの訓練データにおける指の動きの良さについてのスコアを、ギターの教師が与える。スコアを目的変数、各関節のDCT特徴を説明変数として、訓練データに対してサポートベクトル回帰(support vector regression)を行い、回帰式を得る。未知データのDCT特徴をこの回帰式に代入し、スコアが正確に求まるかどうかを実験的に検討し、有効な結果を得た。

  • 超音波画像処理とロボット制御に基づく胎児の3次元計測の研究

    2017  

     View Summary

    近年の医者不足のため、胎児検診の自動化が望まれている。本研究では、超音波画像処理技術とロボット制御技術を組み合わせることにより胎児検診の自動化を目指す。ロボットアームの先端に取り付けられた超音波プローブを、妊婦の腹部表面を走査することにより獲得される超音波画像群から、胎児の頭部等を抽出し、頭囲等を計測するアルゴリズムを明らかにするための検討を行う。本研究では、深層学習を用いて、妊婦の子宮を抽出し、胎児と羊水の範囲を判断する方法を提案し、有効な実験結果を得た。また、超音波画像中の胎児の頭部を楕円で近似して検出する新たな方法を提案し、有効性を実験により検証した。

  • 動画像処理を用いた手術の段階の認識と予測の研究

    2016  

     View Summary

     手術における看護師不足への対策として看護師ロボットの実現が望まれている。我々は手術者を支援する看護師ロボットの実現を目指し、手術野の動画像を分析し、人間の看護師と同様に、自動的に手術の進行や状況を把握する方法の研究を行っている。既に我々は、傷の縫合手術を構成する6つの段階のうち2つの段階(手術者の手の動きに対応)を早期認識(early-recognition)する手法を提案し、有効性の見通しを得ている。本期間は、早期認識の対象を、傷の縫合手術における6つの動作に拡張する検討を行った。我々が既に開発した2つの段階の早期認識法を単純に適用する場合と、6つの動作(段階)の順序関係を利用する場合とを比較検討した。その結果、6つ段階のうちいくつかは順序関係を利用することが有効であるという実験結果が得られた。

  • 動画像処理を用いた手術動作の予測法に関する研究

    2015  

     View Summary

     手術における看護師不足への対策として看護師ロボットの実現が望まれている。本研究では手術者を支援する看護師ロボットの実現を目指し、人間の看護師と同様に、手術野の動画像を分析し、自動的に手術の進行や状況を把握する方法の研究を行っている。既に提案した傷の縫合手術における動作認識手法の認識対象を二つの動作から五つの動作に拡張した[1]。さらに、認識手法の動作が終了する前に予測が可能な方法を提案し、検討した。Siding Windowを利用して連続動画像における動作開始点を抽出し、Ryooらの予測法、動的計画法、ベイズ推定等に基づき、動作の予測を行う[2~4]。本提案手法の有効性を実験により検証した。ただし、前述の予測法において用いる3D-SIFTに基づく特徴量の計算時間について、まだ改善する必要がある。

  • 動画像処理による非剛体の挙動解析の研究

    2014  

     View Summary

    (1)動的動画像中の動物体の抽出・追跡  動的RGBD動画像から、移動物体を検出する方法を検討した.3次元特徴点抽出法とSuperVoxel法による領域分割法とを組み合わせ、移動物体と背景に対応する領域を得る方法を提案した。実験の結果,本提案法の有効性の見通しを得た。 (2)動画像処理に基づく手術の状況解析  研究では、カメラにより手術野を撮像して得られる画像から、手術者の手の動作を解析することにより、縫合手術における「縫合」と「結紮」を認識する方法を提案した。即ち、新たな画像特徴と認識法の組み合わせを提案した。さらに、手術動作の完了前に、その動作を確率的に予測する方法を提案した。これらの手法の有効性を実験的に検証した。

  • 3次元超音波画像と胎児鏡面画像を用いた胎児の口と気道の検出法に関する研究

    2013  

     View Summary

     本研究では、胎児の先天的横隔膜ヘルニアの手術による治療法であるFETO (fetoscopic tracheal occlusion: 胎児鏡を利用する気管の閉塞)に有効と考えられる可撓性のある細い手術具を自律的に胎児の気道にナビゲートし、必要な施術を行う手術法の実現に資するいくつかの医用画像処理技術について検討した。また、内視鏡カメラが装着された手術具を内臓の内部に挿入しながら撮像される動画像を処理し、内臓の内部を3次元復元する方法を検討した。具体的には以下の通りである。(1)楕円の検出 超音波画像中の胎児の頭部に対応する楕円を検出するために、従来の反復的ランダム化ハフ変換を改良した準ランダムな方法を提案し、検討を行った。楕円の一部が遮蔽されている場合でも、従来法より高精度かつ高効率に楕円検出が行えることを実験的に示した。(2)超音波画像からの胎児の顔の3次元検出 3次元超音波画像中において胎児の顔の特徴点(鼻の頂点、両目の目尻、鼻の柱および上唇)間の幾何学的な関係に基づき、胎児の顔を検出するアルゴリズムを提案し、検討を行った。即ち、3次元超音波画像はノイズを多く含むので、3次元画像を構成する2次元のスライス画像ごとに(1)の楕円検出処理を施して胎児の頭部を抽出し、これに基づき顔の輪郭を求め、全てのスライス画像で求まった顔輪郭を統合して顔表面の3次元画像を得る。次に、このようにして得られた3次元顔画像の局所的な曲率と幾何学的な関係を利用して、目、鼻、口等の顔パーツを検出する。胎児ファントムを用いた実験により高い精度で顔の表面と顔パーツが検出できることを示した。(3)内視鏡動画像を用いる内臓内部の3次元復元 胃等の臓器の内部は、皺のような構造が随所に見られるため、内視鏡を動かしながら獲得される動画像からStructure from Motion (SFM)の原理で3次元復元をするための画像特徴点は豊富に存在すると言える。しかし、同じような皺構造の繰り返しが多数見られるため、対応付けが困難であるという課題もある。そこで本研究では、画像の縫い合わせ(stitching)を利用して、動画像におけるフレーム間の対応付けを安定に行う方法を提案した。即ち、SFMにより得られる3次元特徴点集合の縫い合わせを用いて密な3次元復元結果が得られるようにする。実験の結果、正確かつ効率的に3次元復元が行えることが確認できた。

  • 動的動画像中の人物や物体の挙動解析と3次元構造復元の研究

    2013  

     View Summary

     本研究では、移動するKinectセンサにより獲得されるRGB-D(距離)動画像から抽出されたsuper-voxelに基づき、移動物体の追跡を行うとともに、その移動物体の密な三次元構造を復元する方法を検討する。さらに、ビデオ画像から抽出された特徴をベイジアンネットに入力し、人物の無意識な行動を認識する方法を検討する。具体的には以下の通りである。(1)移動RGBD動画像からの移動物体追跡に関する検討移動する Kinect 等のRGBD カメラで複数の移動物体を撮像したシーンの特徴点を背景と各移動物体にクラスタリングする方法には,RANSAC を繰り返し適用する方法等,様々な方法が存在する.Delong らは,マルチラベルのGraphCut を特徴点の分類に用いた手法を提案した。しかし、Delong らの手法では、初期クラスタをランダムに特徴点を3点選択していたため、3点が複数の物体にまたがる場合、処理の精度が低下する問題があった。 そこで本研究では、以下の処理を提案する。(1) RGBD画像をsuper-voxelに分割し、ランダムな3つの特徴点の選択を1つのsuper-voxelに限定して行い、初期クラスタを作成する。(2) 初期クラスタから回転・並進情報を推定する。(3) 各特徴点に対して、推定した回転・並進情報を施すことにより得られる点の位置と、その特徴点におけるオプティカルフローの終点との誤差を計算し、super-voxel内の特徴点の誤差の合計をデータ項とする。(4) Super-voxel間のドロネー図を作成し,隣接するsuper-voxelを結ぶエッジの本数を平滑化項として利用する。(5) グラフカットを行い,ラベルを統合する。グラフカットのエネルギー式は,単一のノード(super-voxel)のみで決まる項であるデータ項と,隣接ノードとの相関によって決まる平滑化項と、ラベルの数とラベル毎のエネルギーの積となる統合処理項の3 つの項で構成される。ラベルの統合の結果によるデータ項・平滑化項のエネルギーの増加よりも、ラベル数の減少によるラベル数の項のエネルギーの減少が大きい場合は統合処理が進む。その結果、ラベルの数が未知の場合でも、適切なラベルの数を自動的に選択することが可能である。 実験を行い、本提案手法の有効性を示す結果を得た。(2)人物の無意識的な行動の認識法の検討 店舗内の顧客の購買行動、特に、迷って/迷わずに、購入したのか/購入しなかったのか、を認識可能な手法を検討した。このような購買行動には定型パターンは乏しいので、ベイジアンネットワークを用いて認識を行う方法を従来から検討している。ただ、従来の検討では、ベイジアンネットワークの各ノードの21個の画像特徴量は、手動で取得していた。良好な認識結果を得ていたものの、手動特徴抽出は実用的とは言えない。そこで、本期間は、画像特徴量を自動的に得る方法を検討した。Felzenszwalb らが提案したDeformable Part Model(DPM)等を利用して、人物が様々な姿勢をとっても追跡が可能な手法を提案し、前述の21個の特徴のうち10個が自動的に計算可能となった。認識実験の結果、若干まだ認識性能が手動特徴の場合より劣るが、有効性の見通しを得たと言える。

  • 動的・静的カメラにより獲得される動画像からの動物体の挙動認識法の研究

    2009  

     View Summary

     本研究では、以下に示す(1)~(4)の課題についてそれぞれ検討を行った。(1)人物全身像の姿勢推定 動作している人物の時々刻々の姿勢を、静止カメラ画像から推定するためには、人体パーツを追跡することが有効である。ここで、パーツが別のパーツを隠すオクルージョンの対処法の確立が重要である。そこで、人体パーツごとに異なった色に着色された衣服を装着した人物を撮像した動画像中で、パーティクルフィルターを用いて人体パーツを追跡する方法を検討した。追跡対象のパーツにオクルージョンが発生したか否かを、パーティクルの挙動により区別できる見通しが得られた。(2)未知人物の行動認識 人物、行動、時系列データから構成されるテンソルをデータベースとして予め構築しておき、テンソル分解を利用して未知人物の行動を認識する方法の検討を進め、我々の方法が関連手法より高い認識率を与える結果が得られた。一方、本手法を利用して、視野が重ならない複数カメラの視野間を移動する人物の対応付けを行うことにより追跡を行う手法を新たに提案し、検討を開始し、有効性の見通しを得た。(3)移動カメラ画像からの移動物体の抽出 パンティルトカメラにより獲得される動画像からカメラの動きを推定し、その動画像中にCG画像を正確に配置できるようにする方法を検討した。特に、動画像中に動く物体を含む場合、その動物体に対応する点からカメラの動きを推定しようとすると、推定精度の低下を招くので、動画像中の動く物体から推定されたか否かを検証する方法を提案し、有効性の見通しを得た。(4)移動カメラ画像からの人物の動き認識 移動ロボットに搭載したカメラにより獲得される動画像から、人物の手の動作を認識する方法の検討を進めた。カメラの動きを考慮に入れて手の軌跡を推定し、これを用いて従来検討したCondensation Algorithmに加え、主成分分析を用いる方法により手話動作の認識の性能評価を行い、主成分分析の方が高い認識率を与える結果を得た。

  • 仮想コミュニケーション環境のための人物像の認識と生成に関する研究

    2000  

     View Summary

     現実には距離を隔てた人物同士の仮想的シーンを介したコミュニケーション環境実現のためには、現実の空間における人物の像を画像処理により自動的に認識し、仮想的シーン中の3次元人物モデルにおいて実時間で再現する、人物像の認識・生成法の実現が必要である。本研究課題では、人物像の認識・生成法の研究を立ち上げるため、以下の項目について、環境整備、方式検討および予備的検討を行ったので、その概要を述べる。(1)人物と物体のインタラクションの認識法の検討 人物のふるまいを画像処理により認識する従来の方式のほとんどは、人物が単独で存在する状況のみを扱っていた。これに対して、現実の空間では、人物はシーンや物体とのインタラクションを日常的に行う。従って、人間にとって自然なマンマシンインタフェースシステムや、自動監視システムを実現するためには、このような人物と物体とのインタラクションを自動的に認識する方式の開発が必要である。そこで、人物のシルエット像の輪郭解析法、移動物体の追跡法、人物と物体が接する場所周辺の濃淡画像解析を統合的に用いる手法を提案し、現在基本的なプログラムを開発中である。(2)3次元顔モデルの生成と表情再現法の検討 従来の3次元顔モデルには、解剖学的な構造を忠実に再現したものはほとんど見られなかった。特に、頭蓋骨、表情筋、脂肪組織、皮膚組織の構造のモデリングは、表情再現のために重要と考えられる。また、表情筋の動作に伴う、皮膚表面の形状変化の再現も重要である。そこで、CT画像データを用いて前述のような顔の構造をモデリングするとともに、実際の人物の表情変化を実現するための表情筋の動的特性パラメータの決定を実写顔画像から行う方法を提案し、現在環境を整備中である。(3)表情再現のリアリティ向上方式の検討 実際の人物の表情を画像処理により推定し、その結果を通信回線を用いて受信側の3次元顔モデルに送信し、表情を再現する方式では、3次元顔モデルにおいて再現される表情のリアリティに課題が残っていた。そこで、実人物の顔表面のテクスチャを顔画像から実時間で獲得し、これを3次元顔モデルにマッピングする方式を提案した。ここで、皮膚表面の光学的特性は個人毎に異なるので、これを効率的に推定する手法を開発する必要がある。また、実人物のいる環境の照明条件を、3次元顔モデルにテクスチャをマッピングする際に反映する必要がある。さらに、3次元顔モデルの形状変化や位置・姿勢変化にも対応する必要がある。現在、これらの検討を可能とする実験環境の整備を進めるとともに、アルゴリズムの詳細を検討中である。

▼display all

Overseas Activities

  • ヒューマノイドロボットの視覚・聴覚・触覚による状況認識とナビゲーション法

    2005.04
    -
    2005.09

    ドイツ   カールスルーエ大学

 

Syllabus

▼display all