• Aucun résultat trouvé

Real-Time Cow Counting From Video or Image Stack captured from a Drone

N/A
N/A
Protected

Academic year: 2021

Partager "Real-Time Cow Counting From Video or Image Stack captured from a Drone"

Copied!
8
0
0

Texte intégral

(1)

HAL Id: hal-03188166

https://hal.archives-ouvertes.fr/hal-03188166

Submitted on 1 Apr 2021

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Real-Time Cow Counting From Video or Image Stack

captured from a Drone

Rutvik Chauhan, Utkarsh Pandey

To cite this version:

Rutvik Chauhan, Utkarsh Pandey. Real-Time Cow Counting From Video or Image Stack captured

from a Drone. [Research Report] University of Alberta. 2021. �hal-03188166�

(2)

Real-Time Cow Counting From Video or Image

Stack captured from a Drone

Rutvik Chauhan

dept. Computer Science University of Alberta

Multimedia Edmonton, Canada rchauha1@ualberta.ca

Utkarsh Pandey

dept. Computer Science University of Alberta

Multimedia Edmonton, Canada upandey@ualberta.ca

Abstract—With the increase in amount of cows in large farm, the demand increases to monitor cows as well as the difficulty to hire a labor to monitor the cows. Due to recent advancements in technology, remote monitoring system can be more reliable using some tags or using computer vision system which can yield more accurate and valuable information about an animal. Also the information about it’s behaviour and surrounding environment can be obtained. Using which we can maximize production.

Index Terms—Object detection, Instance segmentation

I. INTRODUCTION

For large amount of cows and in a large area, counting cows can be tedious task to do for a human.Instead automatic monitoring system can be used for counting cows from video or stack of images. Video or stack of images can be gathered from a drone and we need to detect each cow, color each detected uniquely identified cow and evaluate the performance between video and stack of images.

II. RELATEDWORK

A. Automatic Cow Counting

In [Sha+20], authors used Convolutional Neural Net-work(CNN) to detect and count cattle from Unmanned Aerial Vehicle(UAV) imagery. The algorithm proposed by author takes the following steps: 1) Detect cattle in each image. 2) Reconstruct the 3D surface of the pasture by Structure from Motion(SfM) from the same set of images. With the candidate bounding boxes, the 3D surface, and the relative position of the cameras.3) Merge the per frame detection results guided by the 3D surface. Through this merging over time, the double detection of single cows are eliminated and the correctly counted number of cows is obtained.In [Xu+20], authors used maskrcnn to detect and classify cattle and sheep and to count the total numbers from aerial images taken from quadcopter. Object Detection: In [BWL20], authors achieved 43.5 percent AP (65.7 percent AP50) for the MS COCO dataset at a realtime speed of around 65 FPS on Tesla V100. Instance Seg-mentation: In [He+17], authors proposed MaskRCNN method for object instance segmentation. For instance segmentation [Bol+19a] and [Bol+19b] achieved realtime performance with good accuracy. In [Bol+19a], authors presented a simple, fullyconvolutional model for realtime instance segmentation

that achieves 29.8 mAP on MSCOCO at 33.5 fps evaluated on a single Titan Xp.

1) Implemented algorithms from existing work: For auto-matic monitoring system, cow detection and numbering can be implemented using computer vision or deep learning methods. As seen in literature, traditional computer vision approaches yield less accuracy. To get the state-of-the-art performance, deep learning based methods can be used. For object de-tection and instance segmentation, there are many state-of-the-art models and architectures available.The following was a part of the analysis done by our team. a) Use moving labels/Tags to number the moving cows in a segment of the video (bounding box or segmented cow).b) Segment the cows when they are packed together.c) Label cows entering the view.d) Ensure the label is not reallocated(duplicate) for the cow which re-enters the scene.e) Store unique features and neighbourhood information using a complex data structure to connect two scenes where camera viewpoint changes. The proposed solution can be used in other domains/applications for example- unique vehicle identification, cargo tracking on the docks, and so on. There are certain points that can be considered as challenges involved in this project such as - a) Both the cows and camera(drone) are moving.b) Segment the cows when they are too close to each other.c) Resolution of the camera fit on the drone is not able to display the tag numbers of the cow’s ears.The expected deliverable is feasible if the videos used in the experiments are not overly challenging.Two major milestones that can be considered here would be 1) Counting with cows entering and leaving the view, but without re-entering.

2) Take a short video clip with cow(s) re-appearing in view and introduce additional parameters or data structures to avoid duplicate counting.

B. Review of Other published papers

1) Gradient Vector Flow based Active Shape Model for Lung Field Segmentation in Chest Radiographs: [1]Accurate lung field segmentation is crucial to computer-aided diagnosis (CAD) of lung diseases such as lung cancer and tuberculosis

(3)

Fig. 1. Comparison between yolact, yolo, tensorflow object detection api, amazon rekognition

Fig. 2. Comparison between yolact, yolo, tensorflow object detection api, amazon rekognition

(TB). In this paper, a modified gradient vector flow based active shape model (GVF-ASM) for lung field extraction from chest radiographs is proposed.Experimental results show that the proposed technique provides around 3-5

Interestingly to point out, the proposed solution shows a GVF-ASM method with exponential points evolution for lung field segmentation. The experimental results demonstrated that the proposed method improves the robustness and accuracy of

(4)

the segmentation.

2) Pose Recognition using the Radon Transform: [2]Hu-man pose recognition is an active topic of vision research that has applications in diverse fields such as collaborative virtual environments and robot tele-operation. The proposed paper talks about a novel method for the recognition of human pose using the Radon transform. A binary skeleton representation of the human subject is computed and information concerning the pose is extracted from the parametric Radon transform. The Radon transform generates maxima corresponding to specific orientations of the skeletal representation. A spatial maxima mapping technique is developed to recognize pose from the Radon transform maxima. This technique has the potential to provide information about the exact orientation of pose segments. Experiments with real image data highlight the advantages of this approach.

In this paper, an algorithm for the recognition of human poses and gestures using the Radon transform is proposed. While there are other template-based approaches in contem-porary literature that are being used for pose recognition, these approaches are computationally more expensive and also require normalization of the images. The proposed approach on the other hand does not require any normalization and also guarantees a higher resolution of recognition for arm and leg poses. The promising recognition rates achieved with the parameterized Radon transform emphasize the feasibility of the proposed method. In the future we plan to use the orientation information derived from the Radon transform space to reconstruct motion for virtual models.

3) Omni-directional sensors for pipe inspection: [3]Omni-directional sensors are useful in obtaining a 360” field of view with a single lens camera. At present, for inspecting the interior of pipes (used in industrial processes) fish-eye type lenses are being used. However, fish-eye lenses provide more detail on the empty center part of the pipe and less detail on the surface of the pipe near the camera. This problem can be resolved by using an omni-directional camera, which projects a cylindrical strip of the interior surface of a pipe to a single CCD device. A description of the hardware developed and experimental results are presented.

The use of a conical mirror presents several unusual prob-lems, each of which must be addressed if image quality is to remain high.In theory, the non-planar mirror distortion could be corrected through de-convolution postprocessing. In practice, this approach is problematic; the process is very computationally intensive - especially since the deconvolution basis function is not constant, but is itself a function of the position within the image. This is due to the fact that the radius of curvature changes with distance to the conical apex.

There are several approaches to deal with the low sampling frequencies found at the centre of the CCD : 0 By introducing a curve to the radial profile of the conical mirror, it is possible to obtain complete control over the effective spacial acuity at any given elevation. Factoring in the non-planar distortion, it is then straightforward to control one source of acuity failure with respect to the other,yielding a uniform image

quality across all elevations.e The centre of the cone can be truncated, and conventional wide-angled optics installed within the cone, along its axis (Figure 11). This arrangement can usefully employ the otherwise wasted CCD area by providing a forward view to complement the scope of the reflective cone. An almost complete 4.rr steradians field of view is possible

Colour information can often be represented at spatial frequencies rather lower than those of luminance information (eg. broadcast video standards).Another potential use tion of the cone would be t formation from the inside of the pipeline, using 3 coloured bands applied to the cone. Since the camera will be propelled through the pipe, a fixed point on the surface of the peach distance from the centr is passed by the probe. By monitoring the differential intensities of the light arriving on the (monochrome) CCD as it sweeps through red, green, and blue regions of the cone, postprocessing could reconstruct a full colour representation of the imaged surface. This technique may well prove particularly important, as a feature common to all very wide angle optical systems is that the acuity is very often limited by the CCD devices. Extracting colour from a monochrome device would allow a much higher device resolution.

4) Eye Tracking and Animation for MPEG-4 Coding: [4]Accurate localization and tracking of facial features are cru-cial for developing high quality model-based coding (MPEG-4) systems. For teleconferencing applications at very low bit rates, it is necessary to track eye and lip movements accurately over time. These movements can be coded and transmitted to a remote site, where animation techniques can be used to synthesize facial movements on a model of a face. In this paper we describe simple heuristics which are effective in improving the results of well-known facial feature detection and tracking algorithms. Animation models are also presented, along with experimental results to demonstrate the system being developed. the focus is on the discussion only on the detection, tracking and modeling of eye movements

5) Stereo Matching Using Random Walks: [5]This paper presents a novel two-phase stereo matching algorithm using the random walks framework. At first, a set of reliable matching pixels is extracted with prior matrices defined on the penalties of different disparity configurations and Laplacian matrices defined on the neighbourhood information of pixels. Following this, using the reliable set as seeds, the disparities of unreliable regions are determined by solving a Dirichlet problem. The variance of illumination across different images is taken into account when building the prior matrices and the Laplacian matrices, which improves the accuracy of the resulting disparity maps. Even though random walks have been used in other applications, our work is the first application of random walks in stereo matching. The proposed algorithm demonstrates good performance using the Middlebury stereo datasets.

One crucial and traditional task in stereo vision is stereo matching, the goal of which is to determine the disparities of corresponding pixels in a pair of stereo images.A recent application of random walks (RW) in multilabel image

(5)

seg-mentation shows great potential for its application in stereo matching.The major advantage of random walks is that an exact and unique minimum solution of an energy function of the above form can be produced, while other approaches like GC can only produce an approximation. Therefore, it is very likely that RW can achieve better performance than GC and other optimization frameworks in stereo matching. Hence, we develop a new stereo matching algorithm using random walks. Our contribution lies in that our work is the first application of random walks in the area of stereo matching. Two major phases are considered: first, a set of reliable matching pixels is computed using RWwith prior models (Section 4); second, with the reliable set serving as seeds, the disparities of unreliable pixels are determined using the original RW[4] (Section 5). As demonstrated in Section 6, our algorithm produces more accurate disparity maps than many of GC-based and DP based methods, and some of BP-based methods. In the future, there is a plan to incorporate segmentbased methods to label a more reliable set of initial matching pixels, which will be used to initialize the random walks optimization procedure. This combination, together with GPU implementation, will achieve a much more accurate and faster disparity estimation.

6) Variable-resolution character thinning: [6]Character thinning is often used as a prepossessing step for character recognition. The input to a thinning algorithm is a binary im-age of a character. The pixels with value 1 form the character (object) and the pixels with value 0 form the background. The strokes of the input character are several pixels thick. The objective of thinning is to generate the skeleton of the character in order to achieve information reduction and extraction for further recognition.

The algorithm is parallel in nature and some computation techniques are reported.The result of the proposed method show some significant improvements on several characters in the experiments.More experiments are being conducted to improve the behavior of thinning algorithm on different corners of the character.

7) Automatic Segmentation of Spinal Cord MRI Using Symmetric Boundary Tracing: [7]An adaptive active contour tracing algorithm for extraction of spinal cord from MRI that is fully automatic is developed, unlike existing approaches that need manually chosen seeds. One can accurately extract the target spinal cord and construct the volume of interest to provide visual guidance for strategic rehabilitation surgery planning.

An image of a typical 2-D spinal cord slice is shown in Fig. 1(a). The contrast of the pixels representing the spinal cord in a 2-D slice is distinctly different with respect to the pixel values of the vertebra bone and muscles seen in the image. The segmentation result of the spinal cord is shown in Fig. 1(b). Given that the spinal cord is identified, a technique for the image unskewing process, which is a precursor to the segmentation of other target regions is presented.

The two major proposals of this paper are the symmetry line estimation followed by the active tracing approach. The

Fig. 3.

2-D image slices of spinal cord MRI was acquired with a 2-D spoiled gradient echo sequence operating in multi-slice mode.The proposed paper talks about an automatic spinal cord image segmentation approach in order to assist rehabilitation surgery planning process.Differing from other semiautomatic techniques that extract only the spinal cord, our approach seg-ments the neighboring regions of interest providing better vi-sual guidance for ISMS implant. The methodology introduces a novel application of active tracing of segment boundary,the evolution of which is implemented on a unique contour energy minimization surface.

8) Interactive Multimedia for Adaptive Online Education†: [8]Interactive multimedia content has the capability to im-prove learning performance by enhancing user satisfaction and engagement. Multimedia content also provides better concept representation, which is not possible in conventional multiple-choice and fill-in-the-blank formats.This article, presents a broader view of multimedia education especially for the years to come. The vision is to provide publicly accessible education anywhere, at anytime and to anyone. We propose an innovative CROME (Computer Reinforced Online Multimedia Educa-tion) framework integrating the main components of education including learning, teaching and testing, as well as adaptive testing and student modeling.

The unique contributions of the system are: Automatic difficulty level estimation for selected subjects, compared to estimates based on user group calibration ; Adaptive test-ing ustest-ing innovative item formats versus ustest-ing traditional multiple choice items (Moodle’s multiple choice, GMAT). Computer-assisted automatic question item authoring, rather than manually creating different questions; Custom-designed modules for tracking precise concepts that are lacking in recommendation systems; and, Ability for testing beyond only subject knowledge, unlike other e-learning systems.

(6)

reusability, scalability and interoperability:1) Portability – The implementation is based on Java which is plat-form independent, including J2ME on mobile devices. 2) Reusability – An object-oriented design makes individual components reusable. 3) Interoperability – An authoring tool is an integral part of the CROME system, sharing the item type design, as well as media and item database. 4)Scalability – New item type plug-ins can be added without the need to modify the basic framework or authoring module.

9) Variable Resolution Teleconferencing: [9]Images can contain a large amount of information, which in turns increase the size of the image. To make teleconferencing efficient over bandwidth, some kind of compression technique needs to be implemented. There are mainly 2 types of compression techniques- lossy and lossless. Studies have shown that hu-mans tend to have more focus on the center of the image. Authors have proposed a compression method based on these studies in which the image has higher resolution in the center of the image and lower resolution in the periphery of the image. Authors proposed different variable resolution models. First one was a basic variable resolution model which has 2 parameters- the expected savings (compression) and alpha (which controls the distortion at the edges of the image with respect to the fovea). Second model was Extended model with movable fovea, in which multiple scaling factors are used. Third model was for multiple foveas where two or more regions are displayed with higher resolution than the remainder of the image. Authors listed 3 measures which are important in determining the importance of a compression method: the compression ratio, compression time, resulting image quality. When global image quality is considered, the optimal alpha value differs according to the compression ratio used. On a SUN SpaRC Station SLC, time measured on the CPU for compression using the proposed method was in the range 0.05 to 0.15 seconds. In conclusion, a videoconferencing system is a primary application for virtual resolution as it requires fast compression of images, image quality is not a high priority, typical talking head scene provides for an ideal fovea location. Time of execution of the proposed algorithm made it ideal for the teleconferencing market.

10) A Framework for Adaptive Training Games in Virtual Reality Rehabilitation Environments: [10]For the conven-tional rehabilitation technique and equipment available had very limited time per patient for training. Studies showed that Virtual Reality games engaged patients which leads to longer training time overall. Another limitation is the cost of the electric power wheelchair is high and it was also not flexible in terms of transportation. Specialists available for this type of training were also very limited. Authors proposed a system which had unique features. The system was designed to be low-overhead, portable and focused on indoor training environments. System provides a framework which allows clinicians to design, build and customize interactive indoor rehabilitation training environments which act as a realistic environment. Another challenge is that a simple virtual

re-ality system did not provide any change over time as the patient’s abilities changed. Authors proposed a novel flexible, adaptive framework based on Bayesian Network which allows clinicians to measure performance of the patients and change the difficulty overtime. The system automatically assesses a complex range of patient skills. Evaluation of the system was done with 8 people aged between 22 to 28 year. They are divided in 2 groups: experimental and control. Experiment group has access to a proposed system which enables people to complete challenges in an average time of 81.5 seconds. The control group’s time to complete the challenges were around 104.5 seconds. Rehabilitation specialists and clinicians provided positive feedback for the proposed system.

11) Panoramic Video with Predictive Windows for Telep-resence Applications: [11]Ability to perform tasks in a remote environment is becoming more widely desired in robotics related applications. Telepresence systems provide an immersive impression of the remote site environment to human operators. A typical telepresence system includes a human operator with a control interface, a remote robotic system and a means of communicating between the two systems. Bandwidth for the local and remote systems needs to be sufficient to transmit the sensory and control data. In a limited bandwidth environment, it is very important to decide which fragment of a frame is to be transmitted. Authors proposed using a predictive Kalman filter to model mouse motion in controlling a predictive window. In this paper, author proposed applying the same predictive technique to a panoramic imaging system and consider the problem of parameter identification in the prediction process. Authors also discussed using a Kalman filter to predict viewing direction when displaying panoramic images in a telepresence system. Mouse was used as an input device using which the viewing window around in a panoramic video moved. Operator only saw a fragment of a panoramic image that is contained within this viewing window. Authors’ model assumes that the mouse can be modeled as a point mass under the influence of a constant external force and subject to fractional resistance. The panoramic video system using the proposed method, authors used a camera which had been fitted with a mirror to provide panoramic video. The 360 degree panoramic video allows an immersive feel. The prediction method determines where the mouse is moving so that system can prepare for the operator’s viewing habits to reduce the apparent response delay. Prediction model generally produced a smaller mean distance error than no prediction and the new error modeling improved still further. The research presented used a low resolution imaging system, so overall performance can be increased using higher resolution imaging system.

12) Videoconferencing using Spatially Varying Sensing with Multiple and Moving Foveae: [12]In this paper, a new method for videoconferencing for spatially varying sensing is introduced. Authors proposed a variable resolution system for the compression of the video to transmit it over low bandwidth system. The basic variable resolution model has two parameters- compression and alpha. Basic variable res-olution model is extended to use multiple scaling factors to

(7)

deal with non-rectangular compressed images. Advantage of this method is that the tables need not be recomputed each time the fovea changes the location. In some circumstances there could be more than one area of interest to the observer, which is multiple foveae. There are two distinct approaches for this- cooperative foveae and competitive foveae. In cooperative foveae, it calculates the location of a point in the transformed image with respect to each fovea separately. In competitive foveae, all foveae compete to calculate the position of a point in the transformed image. The fovea which is closest to point in the original image will determines transformed position. Authors also included a videophone prototype using the proposed method which has the interface as shown below in the figure .

Fig. 4. Videophone prototype

13) Nose Shape Estimation Tracking for Model-based Coding: [13]Feature extraction plays a very important role in model-based coding applications and facial recognition. Facial feature extraction is very crucial for facial expression analysis and synthesis, face recognition. In the scenario of Synthetic-Natural Hybrid Coding, extracting and tracking nose shape is very important in generating realistic facial expression. Subtle change in nose shape makes a significant difference in a expression or emotion. Authors proposed a two-stage region growing algorithm- global region growing and local region growing. In global region growing, first the face region is detected and based on luminance and chrominance component of the image, feature regions are extracted. At the end, many different features in terms of blobs will ne detected like mouth, eyes, nose. From these detected blobs top and bottom blobs which contains hair, cloth is removed. Remaining blobs are classified into 4 categories: left eye blob, right eye blob, nose blob, mouth blob. Local region growing algorithm is used to determine local area of features. To detect nostril shape, a geometric template is applied on the nostril region, which is

a twisted pair curve with a leaflike shape. After detecting all the feature regions and estimating shape on various organs, a 3D wireframe model can be matched onto the individual face to track the motion of facial expression. Experimental results are shown in the figure below. Initial results show that this approach is feasible in practical applications.

14) Integrating active face tracking with model based coding: [14]In real situations, detection and tracking of the active talking face is a fundamental step. To compress face-to-face video communication, segmenting and tracking a talking face from a background is a prerequisite. In a movable background, it is difficult to segment the background. In this paper, authors presented a system to track talking faces with an active camera. Proposed system can implement the face detection, tracking, adaptation and animation automatically. To recover the speaker’s silhouette it is necessary to extract it from a number of successive frame differences.

Fig. 5.

To detect eyes and mouth, authors used a deformable template matching. Based on position of the eyes and mouth, position of the head is detected and their shapes can determine the facial expressions. In conclusion, authors proposed a system to track a face with an active camera for an arbitrary background.

15) Modeling Fish-Eye Lenses: [15]We as a human tends to have a variable resolution in our visual system. Foveal image has higher resolution, whereas peripheral information has lower resolution. Existing variable-resolution transforms are very expensive to compute and need special sensors to replicate variable resolution. Authors proposed a Fish Eye Transform which is based on a simplification of the complex logarithmic mapping. Disadvantage of this is that it produces strong anisotropic distortions in peripheral regions. Authors also discussed polynomial fish eye transform to eliminate this problem. Authors also proposed fish eye lenses, using which the objective became to find a function that represents the average distortion calculated. In experiments, the polynomial model appeared to have a better performance in compensating the distortions.

REFERENCES

[1] T. Xu, M. Mandal, R. Long, and A. Basu, “Gradient vector flow based active shape model for lung field segmentation in chest radiographs,” in 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2009, pp. 3561–3564.

[2] M. Singh, M. Mandal, and A. Basu, “Pose recognition using the radon transform,” in 48th Midwest Symposium on Circuits and Systems, 2005. IEEE, 2005, pp. 1091–1094.

[3] A. Basu and D. Southwell, “Omni-directional sensors for pipe inspec-tion,” in 1995 IEEE International Conference on Systems, Man and Cybernetics. Intelligent Systems for the 21st Century, vol. 4. IEEE, 1995, pp. 3107–3112.

(8)

[4] S. Bernogger, A. Basu, L. Yin, and A. Pinz, “Eye tracking and animation for mpeg-4 coding,” in Pattern Recognition, International Conference on, vol. 2. Los Alamitos, CA, USA: IEEE Computer Society, aug 1998, p. 1281. [Online]. Available: https://doi.ieeecomputersociety.org/10.1109/ICPR.1998.711935 [5] R. Shen, I. Cheng, X. Li, and A. Basu, “Stereo matching using random

walks,” in 2008 19th International Conference on Pattern Recognition. IEEE, 2008, pp. 1–4.

[6] X. Li and A. Basu, “Variable-resolution charac-ter thinning,” Pattern Recognition Letters, vol. 12, no. 4, pp. 241–248, 1991. [Online]. Available: https://www.sciencedirect.com/science/article/pii/016786559190038N [7] D. P. Mukherjee, I. Cheng, N. Ray, V. Mushahwar, M. Lebel, and

A. Basu, “Automatic segmentation of spinal cord mri using symmetric boundary tracing,” IEEE Transactions on Information Technology in Biomedicine, vol. 14, no. 5, pp. 1275–1278, 2010.

[8] I. Cheng, A. Basu, and R. Goebel, “Interactive multimedia for adaptive online education.” IEEE Multim., vol. 16, no. 1, pp. 16–25, 2009. [9] A. Basu, A. Sullivan, and K. Wiebe, “Variable resolution

teleconferenc-ing,” in Proceedings of IEEE Systems Man and Cybernetics Conference-SMC, vol. 4. IEEE, 1993, pp. 170–175.

[10] N. Rossol, I. Cheng, W. F. Bischof, and A. Basu, “A framework for adaptive training and games in virtual reality rehabilitation environments,” in Proceedings of the 10th International Conference on Virtual Reality Continuum and Its Applications in Industry, ser. VRCAI ’11. New York, NY, USA: Association for Computing Machinery, 2011, p. 343–346. [Online]. Available: https://doi.org/10.1145/2087756.2087810

[11] J. Baldwin, A. Basu, and H. Zhang, “Panoramic video with pre-dictive windows for telepresence applications,” in Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C), vol. 3, 1999, pp. 1922–1927 vol.3.

[12] A. Basu and K. J. Wiebe, “Videoconferencing using spatially varying sensing with multiple and moving foveae,” in Proceedings of the 12th IAPR International Conference on Pattern Recognition, Vol. 2 -Conference B: Computer Vision Image Processing. (Cat. No.94CH3440-5), 1994, pp. 30–34 vol.3.

[13] Lijun Yin and A. Basu, “Nose shape estimation and tracking for model-based coding,” in 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221), vol. 3, 2001, pp. 1477–1480 vol.3.

[14] L. Yin and A. Basu, “Integrating active face tracking with model based coding,” Pattern Recognition Letters, vol. 20, no. 6, pp. 651–657, 1999. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S016786559900029X [15] W. Shao, R. Kawakami, R. Yoshihashi, S. You, H. Kawase, and

T. Naemura, “Cattle detection and counting in uav images based on convolutional neural networks,” International Journal of Remote Sensing, vol. 41, no. 1, pp. 31–52, 2020. [Online]. Available: https://doi.org/10.1080/01431161.2019.1624858

Figure

Fig. 2. Comparison between yolact, yolo, tensorflow object detection api, amazon rekognition
Fig. 4. Videophone prototype

Références

Documents relatifs

LiveCity e-Health for glaucoma is a part of this European program, which aims to improve the quality of life of chronic patients, reduce their need for transportation to the

The LiveCity project aims on creating a set of plugins that integrate high quality video streaming that is used for the communication between patients and health care

A traditional method for real-time segmentation of moving objet in video based system is “background subtraction”, or computing the error between a background model and

The region segmentation algorithm can be exploited directly using regions boundaries and region color properties or as a source of information on the image content (level of

One key property TADF molecules should fulfill for efficient RISC is a small energy gap between the lowest singlet and triplet excited states ΔE ST (usually < 0.2 eV), so that

Starting from the above caml specication and C code, SKiPPER has been used both to check the correctness of the parallelisation process (by using the sequential emulation

Our approach includes three main steps: a first step of partitioning the original mesh, generating several fragments of the original mesh that can be supported by

Motion estimation is a real bottleneck for real-time implementation, so we rather propose to combine an improved Change De- tection Mask (CDM ) with spatial segmentation in order